ML Platform Podcast cover image

Embracing Responsible AI for ML Models in Production with Amber Roberts

ML Platform Podcast

00:00

Exploring the Relationship between Post-deployments Monitoring and Model Explainability for Responsible AI Models

This chapter explores the importance of tools like drift detection, data quality checks, production checks, and fairness checks in understanding how a model is performing and reacting to new data. It emphasizes the need for observability and highlights the long-term goal of building responsible AI models.

Play episode from 05:04
Transcript

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app