Conversation Evals
User Satisfaction
Asseses the user satisfaction with the conversation
Conversation Satisfaction measures how well a model/LLM answers the query asked by the user.
It determines the user’s satisfaction based on the conversation with the LLM/AI assistant.
Measures the user’s satisfaction with the conversation with the LLM/AI assistant based on completeness and user’s acceptance.
Columns required:
user_persona
: The persona of the user asking the queriesllm_persona
: The persona of the LLM/AI assistant
How to use it?
By default, we are using GPT 3.5 Turbo. If you want to use a different model, check out this tutorial.
Sample Response:
A higher conversation satisfaction score reflects that the user seems to be satisfied with the conversation.
The conversation that the agent was not able to address the user’s query, also indicated by the user’s response: “You don’t understand”.
Resulting in a low conversation satisfaction score.
How it works?
We evaluate conversation satisfaction by determining which of the following three cases apply for the given task data:
- The user looks highly satisifed in the conversation.
- The user looks moderately satisfied in the conversation.
- The user is not at all satisfied in the conversation.
Was this page helpful?