Introduction
This workshop contains examples of real-world scenarios for evaluating generative AI outputs for quality and safety. Your goal is to complete the exercises in the RAI-workshops repository which accompanies this module.
Learning objectives
By completing this module, you're able to:
- Assess a generative AI app response using performance and quality metrics
- Assess a generative AI app response using risk and safety metrics
- Run an evaluation and track the results in Azure AI Foundry portal
- Create a custom evaluator with Prompty
- Send queries to an endpoint and run evaluators on the resulting query and response
- Generate a synthetic dataset using conversation starters
Prerequisites
- An Azure subscription – Create one for free
- Familiarity with Azure and the Azure portal
- Ability to understand Python at a beginner level
- Developing in browser:
- GitHub account
- Developing locally:
- Docker desktop
- Visual Studio Code