Interpreting ML models with explainable AI

We often trust our high-accuracy ML models to make decisions for our users, but it’s hard to know exactly why or how these models came to specific conclusions. Explainable AI provides a suite of tools to help you interpret your ML model’s predictions. Listen to this discussion regarding how to use Explainable AI to ensure our ML models are treating all users fairly. Watch for a presentation on how to analyze image, text, and tabular models from a fairness perspective, using Explanations on AI Platform. Finally, learn how to use the What-if Tool, an open source visualization tool for optimizing your ML model’s performance and fairness.

Speaker: Sara Robinson

Watch more:
Google Cloud Next ’20: OnAir →

Subscribe to the GCP Channel →



product: Cloud – General; fullname: Sara Robinson; event: Google Cloud Next 2020;

Duration: 00:21:08
Publisher: Google Cloud
You can watch this video also at the source.