Artwork

תוכן מסופק על ידי Brian T. O’Neill from Designing for Analytics. כל תוכן הפודקאסטים כולל פרקים, גרפיקה ותיאורי פודקאסטים מועלים ומסופקים ישירות על ידי Brian T. O’Neill from Designing for Analytics או שותף פלטפורמת הפודקאסט שלהם. אם אתה מאמין שמישהו משתמש ביצירה שלך המוגנת בזכויות יוצרים ללא רשותך, אתה יכול לעקוב אחר התהליך המתואר כאן https://he.player.fm/legal.
Player FM - אפליקציית פודקאסט
התחל במצב לא מקוון עם האפליקציה Player FM !

175 - The MIRRR UX Framework for Designing Trustworthy Agentic AI Applications (Part 1)

28:51
 
שתפו
 

Manage episode 498533765 series 2938687
תוכן מסופק על ידי Brian T. O’Neill from Designing for Analytics. כל תוכן הפודקאסטים כולל פרקים, גרפיקה ותיאורי פודקאסטים מועלים ומסופקים ישירות על ידי Brian T. O’Neill from Designing for Analytics או שותף פלטפורמת הפודקאסט שלהם. אם אתה מאמין שמישהו משתמש ביצירה שלך המוגנת בזכויות יוצרים ללא רשותך, אתה יכול לעקוב אחר התהליך המתואר כאן https://he.player.fm/legal.

In this episode of Experiencing Data, I introduce part 1 of my new MIRRR UX framework for designing trustworthy agentic AI applications—you know, the kind that might actually get used and have the opportunity to create the desired business value everyone seeks! One of the biggest challenges with both traditional analytics, ML, and now, LLM-driven AI agents, is getting end users and stakeholders to trust and utilize these data products—especially if we’re asking humans in the loop to make changes to their behavior or ways of working.

In this episode, I challenge the idea that software UIs will vanish with the rise of AI-based automation. In fact, the MIRRR framework is based on the idea that AI agents should be “in the human loop,” and a control surface (user interface) may in many situations be essential to ensure any automated workers engender trust with their human overlords.

By properly considering the control and oversight that end users and stakeholders need, you can enable the business value and UX outcomes that your paying customers, stakeholders, and application users seek from agentic AI.

Using use cases from insurance claims processing, in this episode, I introduce the first two of five control points in the MIRRR framework—Monitor and Interrupt. These control points represent core actions that define how AI agents often should operate and interact within human systems:

  • Monitor – enabling appropriate transparency into AI agent behavior and performance
  • Interrupt – designing both manual and automated pausing mechanisms to ensure human oversight remains possible when needed

…and in a couple weeks, stay tuned for part 2 where I’ll wrap up this first version of my MIRRR framework.

Highlights / Skip to:
  • 00:34 Introducing the MIRRR UX Framework for designing trustworthy agentic AI Applications.
  • 01:27 The importance of trust in AI systems and how it is linked to user adoption
  • 03:06 Cultural shifts, AI hype, and growing AI skepticism
  • 04:13 Human centered design practices for agentic AI
  • 06:48 I discuss how understanding your users’ needs does not change with agentic AI, and that trust in agentic applications has direct ties to user adoption and value creation
  • 11:32 Measuring success of agentic applications with UX outcomes
  • 15:26 Introducing the first two of five MIRRR framework control points:
    • 16:29 M is for Monitor; understanding the agent’s “performance,” and the right
      level of transparency end users need, from individual tasks to aggregate views
    • 20:29 I is for Interrupt; when and why users may need to stop the agent—and
      what happens next
  • 28:02 Conclusion and next steps
  continue reading

106 פרקים

Artwork
iconשתפו
 
Manage episode 498533765 series 2938687
תוכן מסופק על ידי Brian T. O’Neill from Designing for Analytics. כל תוכן הפודקאסטים כולל פרקים, גרפיקה ותיאורי פודקאסטים מועלים ומסופקים ישירות על ידי Brian T. O’Neill from Designing for Analytics או שותף פלטפורמת הפודקאסט שלהם. אם אתה מאמין שמישהו משתמש ביצירה שלך המוגנת בזכויות יוצרים ללא רשותך, אתה יכול לעקוב אחר התהליך המתואר כאן https://he.player.fm/legal.

In this episode of Experiencing Data, I introduce part 1 of my new MIRRR UX framework for designing trustworthy agentic AI applications—you know, the kind that might actually get used and have the opportunity to create the desired business value everyone seeks! One of the biggest challenges with both traditional analytics, ML, and now, LLM-driven AI agents, is getting end users and stakeholders to trust and utilize these data products—especially if we’re asking humans in the loop to make changes to their behavior or ways of working.

In this episode, I challenge the idea that software UIs will vanish with the rise of AI-based automation. In fact, the MIRRR framework is based on the idea that AI agents should be “in the human loop,” and a control surface (user interface) may in many situations be essential to ensure any automated workers engender trust with their human overlords.

By properly considering the control and oversight that end users and stakeholders need, you can enable the business value and UX outcomes that your paying customers, stakeholders, and application users seek from agentic AI.

Using use cases from insurance claims processing, in this episode, I introduce the first two of five control points in the MIRRR framework—Monitor and Interrupt. These control points represent core actions that define how AI agents often should operate and interact within human systems:

  • Monitor – enabling appropriate transparency into AI agent behavior and performance
  • Interrupt – designing both manual and automated pausing mechanisms to ensure human oversight remains possible when needed

…and in a couple weeks, stay tuned for part 2 where I’ll wrap up this first version of my MIRRR framework.

Highlights / Skip to:
  • 00:34 Introducing the MIRRR UX Framework for designing trustworthy agentic AI Applications.
  • 01:27 The importance of trust in AI systems and how it is linked to user adoption
  • 03:06 Cultural shifts, AI hype, and growing AI skepticism
  • 04:13 Human centered design practices for agentic AI
  • 06:48 I discuss how understanding your users’ needs does not change with agentic AI, and that trust in agentic applications has direct ties to user adoption and value creation
  • 11:32 Measuring success of agentic applications with UX outcomes
  • 15:26 Introducing the first two of five MIRRR framework control points:
    • 16:29 M is for Monitor; understanding the agent’s “performance,” and the right
      level of transparency end users need, from individual tasks to aggregate views
    • 20:29 I is for Interrupt; when and why users may need to stop the agent—and
      what happens next
  • 28:02 Conclusion and next steps
  continue reading

106 פרקים

כל הפרקים

×
 
Loading …

ברוכים הבאים אל Player FM!

Player FM סורק את האינטרנט עבור פודקאסטים באיכות גבוהה בשבילכם כדי שתהנו מהם כרגע. זה יישום הפודקאסט הטוב ביותר והוא עובד על אנדרואיד, iPhone ואינטרנט. הירשמו לסנכרון מנויים במכשירים שונים.

 

מדריך עזר מהיר

האזן לתוכנית הזו בזמן שאתה חוקר
הפעלה