Azure AI Foundry ValueError running evaluator (coherence, fluency) ValueError caused by unexpected evaluation LLM outputs. #39011
Labels
customer-reported
Issues that are reported by GitHub users external to the Azure organization.
Evaluation
Issues related to the client library for Azure AI Evaluation
needs-team-attention
Workflow: This issue needs attention from Azure service team or SDK team
question
The issue doesn't require a change to the product in order to be resolved. Most issues start as that
Service Attention
Workflow: This issue is responsible by Azure service team.
Describe the bug
When completing the Evaluate the performance of generative AI apps with Azure AI Foundry course the automated Coherence and Fluency evaluations fail due to ValueErrors.
To Reproduce
Steps to reproduce the behavior:
Expected behavior
A clear and concise description of what you expected to happen.
Expect for the evaluation LLMs to provide a consistent output and have a common util which can effectively catch edge cases.
Screenshots
If applicable, add screenshots to help explain your problem.
Additional context
Could be improved by updating/re-evaluating the few shot prompts (possibly using the promptflow style or even add Chain of Thought. Also could change parse_quality_evaluator_reason_score function to pull out just the number for the scoring component.
The text was updated successfully, but these errors were encountered: