Episode

Automated safety evaluations for generative AI applications

with Seth Juarez, Minsoo Thigpen

In this episode, we learn more about new metrics for automated safety evaluations in Azure AI Studio. These evaluations make it easier for organizations to systematically and comprehensively test the quality, safety, and reliability of their LLM application before deploying it in production.

Chapters

  • 00:00 - Introduction
  • 00:20 - On today's episode
  • 01:15 - Automated safety evaluations
  • 06:54 - Evaluations in VS Code demonstration
  • 13:54 - Evaluations in Azure AI Studio demonstration
  • 17:10 - Metric dashboard
  • 22:00 - Learn more

Connect

Azure
Azure OpenAI Service