scispace - formally typeset
Open AccessJournal ArticleDOI

"Hello AI": Uncovering the Onboarding Needs of Medical Practitioners for Human-AI Collaborative Decision-Making

TLDR
This work investigates the key types of information medical experts desire when they are first introduced to a diagnostic AI assistant, providing a richer understanding of what experts find important in their introduction to AI assistants before integrating them into routine practice.
Abstract
Although rapid advances in machine learning have made it increasingly applicable to expert decision-making, the delivery of accurate algorithmic predictions alone is insufficient for effective human-AI collaboration. In this work, we investigate the key types of information medical experts desire when they are first introduced to a diagnostic AI assistant. In a qualitative lab study, we interviewed 21 pathologists before, during, and after being presented deep neural network (DNN) predictions for prostate cancer diagnosis, to learn the types of information that they desired about the AI assistant. Our findings reveal that, far beyond understanding the local, case-specific reasoning behind any model decision, clinicians desired upfront information about basic, global properties of the model, such as its known strengths and limitations, its subjective point-of-view, and its overall design objective--what it's designed to be optimized for. Participants compared these information needs to the collaborative mental models they develop of their medical colleagues when seeking a second opinion: the medical perspectives and standards that those colleagues embody, and the compatibility of those perspectives with their own diagnostic patterns. These findings broaden and enrich discussions surrounding AI transparency for collaborative decision-making, providing a richer understanding of what experts find important in their introduction to AI assistants before integrating them into routine practice.

read more

Citations
More filters
Proceedings ArticleDOI

Interpreting Interpretability: Understanding Data Scientists' Use of Interpretability Tools for Machine Learning

TL;DR: It is indicated that data scientists over-trust and misuse interpretability tools, and few of their participants were able to accurately describe the visualizations output by these tools.
Proceedings ArticleDOI

A Human-Centered Evaluation of a Deep Learning System Deployed in Clinics for the Detection of Diabetic Retinopathy

TL;DR: A human-centered study of a deep learning system used in clinics for the detection of diabetic eye disease in Thailand indicates that several socio-environmental factors impact model performance, nursing workflows, and the patient experience.
Journal ArticleDOI

How Machine Learning Will Transform Biomedicine

TL;DR: A vision for how machine learning can transform three broad areas of biomedicine: clinical diagnostics, precision treatments, and health monitoring, where the goal is to maintain health through a range of diseases and the normal aging process is outlined.
Posted Content

Does the Whole Exceed its Parts? The Effect of AI Explanations on Complementary Team Performance

TL;DR: This work conducts mixed-method user studies on three datasets, where an AI with accuracy comparable to humans helps participants solve a task (explaining itself in some conditions), and observes complementary improvements from AI augmentation that were not increased by explanations.
References
More filters
Journal ArticleDOI

Using thematic analysis in psychology

TL;DR: Thematic analysis is a poorly demarcated, rarely acknowledged, yet widely used qualitative analytic method within psychology as mentioned in this paper, and it offers an accessible and theoretically flexible approach to analysing qualitative data.
Proceedings ArticleDOI

"Why Should I Trust You?": Explaining the Predictions of Any Classifier

TL;DR: In this article, the authors propose LIME, a method to explain models by presenting representative individual predictions and their explanations in a non-redundant way, framing the task as a submodular optimization problem.
Journal ArticleDOI

Dermatologist-level classification of skin cancer with deep neural networks

TL;DR: This work demonstrates an artificial intelligence capable of classifying skin cancer with a level of competence comparable to dermatologists, trained end-to-end from images directly, using only pixels and disease labels as inputs.
Journal ArticleDOI

What can "thematic analysis" offer health and wellbeing researchers?

TL;DR: The field of health and wellbeing scholarship has a strong tradition of qualitative research* and rightly so, and rich and compelling insights into the real worlds, experiences, and perspectives of patients and health care professionals can be found through quantitative methods.
Related Papers (5)
Trending Questions (1)
What are the needs of ai?

The needs of AI include understanding its strengths and limitations, its subjective point-of-view, and its overall design objective.