We often trust our high-accuracy ML models to make decisions for our users, but it’s hard to know exactly why or how these models came to specific conclusions. Explainable AI provides a suite of tools to help you interpret your ML model’s predictions. Listen to this discussion regarding how to use Explainable AI to ensure our ML models are treating all users fairly. Watch for a presentation on how to analyze image, text, and tabular models from a fairness perspective, using Explanations on AI Platform. Finally, learn how to use the What-if Tool, an open source visualization tool for optimizing your ML model’s performance and fairness.
Speaker: Sara Robinson
Google Cloud Next ’20: OnAir → https://goo.gle/next2020
Subscribe to the GCP Channel → https://goo.gle/GCP
product: Cloud – General; fullname: Sara Robinson; event: Google Cloud Next 2020;
Publisher: Google Cloud
You can watch this video also at the source.