Usage and Examples
This page provides practical guidance on how to use the Azure AI Evaluation SDK through hands-on examples and common evaluation scenarios. Itβs designed to help users quickly get started and build confidence in applying the SDK to real-world tasks.
1. Basic Evaluation Workflow
- Step-by-step guide:
- Load dataset
- Configure evaluation pipeline
- Run evaluation
- View results
- π Quickstart: Evaluate a Model Locally
2. Sample Code Snippets
- Text Classification Example:
- RAG Evaluation Example:
π Evaluating RAG Applications
3. Configuration Options
- YAML Config Structure:
- Define model endpoints
- Specify metrics
- Set dataset paths
- π Configuration Reference
4. Common Scenarios
- Evaluating Azure OpenAI Models
- Comparing Multiple Models
- Evaluating Fairness Across Demographics
- π Fairness Evaluation Guide
5. Troubleshooting Tips
- Common Errors:
- Invalid config paths
- Unsupported dataset formats
- Metric computation failures
- Debugging Tools:
- Verbose logging
- Evaluation reports
- π Troubleshooting Evaluation SDK