Response Quality Evals
Response Conciseness
Grades how concise the generated response is or if it has any additional irrelevant information for the question asked.
Response conciseness score measures whether the generated response contains any additional information irrelevant to the question asked.
Columns required:
question
: The question asked by the userresponse
: The response given by the model
How to use it?
By default, we are using GPT 3.5 Turbo for evaluations. If you want to use a different model, check out this tutorial.
Sample Response:
A higher response conciseness score reflects that the response is concise and does not contain any irrelevant information.
The response has information about software integration, architectural patterns, and design principles.
This information is not relevant to the user’s question, “What are the primary components of a cell?”
Resulting in a low response conciseness score.
How it works?
We evaluate response conciseness by determining which of the following three cases apply for the given task data:
- The generated answer has a lot of additional irrelevant information.
- The generated answer has a little additional irrelevant information.
- The generated answer has no additional irrelevant information.
Was this page helpful?