Episode
Automated safety evaluations for generative AI applications
with Seth Juarez, Minsoo Thigpen
In this episode, we learn more about new metrics for automated safety evaluations in Azure AI Studio. These evaluations make it easier for organizations to systematically and comprehensively test the quality, safety, and reliability of their LLM application before deploying it in production.
Chapters
- 00:00 - Introduction
- 00:20 - On today's episode
- 01:15 - Automated safety evaluations
- 06:54 - Evaluations in VS Code demonstration
- 13:54 - Evaluations in Azure AI Studio demonstration
- 17:10 - Metric dashboard
- 22:00 - Learn more
Recommended resources
Related episodes
- Join us every other Friday for an AI Show livestream on YouTube
- Subscribe to the AI Show
- AI Show Playlist
Connect
- Seth Juarez | Twitter: @sethjuarez
In this episode, we learn more about new metrics for automated safety evaluations in Azure AI Studio. These evaluations make it easier for organizations to systematically and comprehensively test the quality, safety, and reliability of their LLM application before deploying it in production.
Chapters
- 00:00 - Introduction
- 00:20 - On today's episode
- 01:15 - Automated safety evaluations
- 06:54 - Evaluations in VS Code demonstration
- 13:54 - Evaluations in Azure AI Studio demonstration
- 17:10 - Metric dashboard
- 22:00 - Learn more
Recommended resources
Related episodes
- Join us every other Friday for an AI Show livestream on YouTube
- Subscribe to the AI Show
- AI Show Playlist
Connect
- Seth Juarez | Twitter: @sethjuarez
Have feedback? Submit an issue here.