התחל במצב לא מקוון עם האפליקציה Player FM !
פודקאסטים ששווה להאזין
בחסות


1 How To Replace A $100,000+ Salary Within 6 MONTHS Through Buying A Small Business w/ Alex Kamenca & Carley Mitus 57:50
Experiencing Data w/ Brian T. O’Neill (UX for AI Data Products, SAAS Analytics, Data Product Management)
«
»
125 - Human-Centered XAI: Moving from Algorithms to Explainable ML UX with Microsoft Researcher Vera Liao
Manage episode 376109532 series 2527129
Today I’m joined by Vera Liao, Principal Researcher at Microsoft. Vera is a part of the FATE (Fairness, Accountability, Transparency, and Ethics of AI) group, and her research centers around the ethics, explainability, and interpretability of AI products. She is particularly focused on how designers design for explainability. Throughout our conversation, we focus on the importance of taking a human-centered approach to rendering model explainability within a UI, and why incorporating users during the design process informs the data science work and leads to better outcomes. Vera also shares some research on why example-based explanations tend to out-perform [model] feature-based explanations, and why traditional XAI methods LIME and SHAP aren’t the solution to every explainability problem a user may have.
Highlights/ Skip to:
- I introduce Vera, who is Principal Researcher at Microsoft and whose research mainly focuses on the ethics, explainability, and interpretability of AI (00:35)
- Vera expands on her view that explainability should be at the core of ML applications (02:36)
- An example of the non-human approach to explainability that Vera is advocating against (05:35)
- Vera shares where practitioners can start the process of responsible AI (09:32)
- Why Vera advocates for doing qualitative research in tandem with model work in order to improve outcomes (13:51)
- I summarize the slides I saw in Vera’s deck on Human-Centered XAI and Vera expands on my understanding (16:06)
- Vera’s success criteria for explainability (19:45)
- The various applications of AI explainability that Vera has seen evolve over the years (21:52)
- Why Vera is a proponent of example-based explanations over model feature ones (26:15)
- Strategies Vera recommends for getting feedback from users to determine what the right explainability experience might be (32:07)
- The research trends Vera would most like to see technical practitioners apply to their work (36:47)
- Summary of the four-step process Vera outlines for Question-Driven XAI design (39:14)
Links
113 פרקים
Manage episode 376109532 series 2527129
Today I’m joined by Vera Liao, Principal Researcher at Microsoft. Vera is a part of the FATE (Fairness, Accountability, Transparency, and Ethics of AI) group, and her research centers around the ethics, explainability, and interpretability of AI products. She is particularly focused on how designers design for explainability. Throughout our conversation, we focus on the importance of taking a human-centered approach to rendering model explainability within a UI, and why incorporating users during the design process informs the data science work and leads to better outcomes. Vera also shares some research on why example-based explanations tend to out-perform [model] feature-based explanations, and why traditional XAI methods LIME and SHAP aren’t the solution to every explainability problem a user may have.
Highlights/ Skip to:
- I introduce Vera, who is Principal Researcher at Microsoft and whose research mainly focuses on the ethics, explainability, and interpretability of AI (00:35)
- Vera expands on her view that explainability should be at the core of ML applications (02:36)
- An example of the non-human approach to explainability that Vera is advocating against (05:35)
- Vera shares where practitioners can start the process of responsible AI (09:32)
- Why Vera advocates for doing qualitative research in tandem with model work in order to improve outcomes (13:51)
- I summarize the slides I saw in Vera’s deck on Human-Centered XAI and Vera expands on my understanding (16:06)
- Vera’s success criteria for explainability (19:45)
- The various applications of AI explainability that Vera has seen evolve over the years (21:52)
- Why Vera is a proponent of example-based explanations over model feature ones (26:15)
- Strategies Vera recommends for getting feedback from users to determine what the right explainability experience might be (32:07)
- The research trends Vera would most like to see technical practitioners apply to their work (36:47)
- Summary of the four-step process Vera outlines for Question-Driven XAI design (39:14)
Links
113 פרקים
すべてのエピソード
×
1 167 - AI Product Management and Design: How Natalia Andreyeva and Team at Infor Nexus Create B2B Data Products that Customers Value 37:34

1 166 - Can UX Quality Metrics Increase Your Data Product's Business Value and Adoption? 26:12

1 165 - How to Accommodate Multiple User Types and Needs in B2B Analytics and AI Products When You Lack UX Resources 49:04

1 164 - The Hidden UX Taxes that AI and LLM Features Impose on B2B Customers Without Your Knowledge 45:25

1 163 - It’s Not a Math Problem: How to Quantify the Value of Your Enterprise Data Products or Your Data Product Management Function 41:41

1 162 - Beyond UI: Designing User Experiences for LLM and GenAI-Based Products 42:07

1 161 - Designing and Selling Enterprise AI Products [Worth Paying For] 34:00

1 160 - Leading Product Through a Merger/Acquisition: Lessons from The Predictive Index’s CPO Adam Berke 42:10

1 159 - Uncorking Customer Insights: How Data Products Revealed Hidden Gems in Liquor & Hospitality Retail 40:47

1 158 - From Resistance to Reliance: Designing Data Products for Non-Believers with Anna Jacobson of Operator Collective 43:41

1 157 - How this materials science SAAS company brings PM+UX+data science together to help materials scientists accelerate R&D 34:58

1 156-The Challenges of Bringing UX Design and Data Science Together to Make Successful Pharma Data Products with Jeremy Forman 41:37

1 155 - Understanding Human Engagement Risk When Designing AI and GenAI User Experiences 55:33

1 154 - 10 Things Founders of B2B SAAS Analytics and AI Startups Get Wrong About DIY Product and UI/UX Design 44:47

1 153 - What Impressed Me About How John Felushko Does Product and UX at the Analytics SAAS Company, LabStats 57:31
ברוכים הבאים אל Player FM!
Player FM סורק את האינטרנט עבור פודקאסטים באיכות גבוהה בשבילכם כדי שתהנו מהם כרגע. זה יישום הפודקאסט הטוב ביותר והוא עובד על אנדרואיד, iPhone ואינטרנט. הירשמו לסנכרון מנויים במכשירים שונים.