Helicone
Helicone helps you understand how your application is performing with its monitoring tools. We will walk you through the use of Helicone for monitoring and log your UpTrain evaluations in Helicone Dashboards
How to integrate?
Prerequisites
Define OpenAI client
Let's define our dataset
Define your prompt
Define funtion to generate responses
Define UpTrain Function to run Evaluations
We have used the following 5 metrics from UpTrain’s library:
-
Response Conciseness: Evaluates how concise the generated response is or if it has any additional irrelevant information for the question asked.
-
Factual Accuracy: Evaluates whether the response generated is factually correct and grounded by the provided context.
-
Context Utilization: Evaluates how complete the generated response is for the question specified given the information provided in the context. Also known as Reponse Completeness wrt context
-
Response Relevance: Evaluates how relevant the generated response was to the question specified.
Each score has a value between 0 and 1.
You can look at the complete list of UpTrain’s supported metrics here
Run the evaluations and log the data to Helicone t
Visualize Results in Helicone Dashboards
You can log into Helicone Dashoards to observe your LLM applications over cost, tokens, latency
You can also look at individual records