scispace - formally typeset
Proceedings ArticleDOI

Formalizing Trust in Artificial Intelligence: Prerequisites, Causes and Goals of Human Trust in AI

TLDR
In this paper, the authors discuss a model of trust inspired by sociologists' notion of interpersonal trust (i.e., trust between people) and discuss how to design trustworthy AI, how to evaluate whether trust has manifested, and whether it is warranted.
Abstract
Trust is a central component of the interaction between people and AI, in that 'incorrect' levels of trust may cause misuse, abuse or disuse of the technology. But what, precisely, is the nature of trust in AI? What are the prerequisites and goals of the cognitive mechanism of trust, and how can we promote them, or assess whether they are being satisfied in a given interaction? This work aims to answer these questions. We discuss a model of trust inspired by, but not identical to, interpersonal trust (i.e., trust between people) as defined by sociologists. This model rests on two key properties: the vulnerability of the user; and the ability to anticipate the impact of the AI model's decisions. We incorporate a formalization of 'contractual trust', such that trust between a user and an AI model is trust that some implicit or explicit contract will hold, and a formalization of 'trustworthiness' (that detaches from the notion of trustworthiness in sociology), and with it concepts of 'warranted' and 'unwarranted' trust. We present the possible causes of warranted trust as intrinsic reasoning and extrinsic behavior, and discuss how to design trustworthy AI, how to evaluate whether trust has manifested, and whether it is warranted. Finally, we elucidate the connection between trust and XAI using our formalization.

read more

Citations
More filters

A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle

TL;DR: In this paper, the authors identify seven potential sources of downstream harm in machine learning, spanning data collection, development, and deployment, and propose a framework to facilitate more productive and precise communication around these issues.
Posted Content

On the Opportunities and Risks of Foundation Models.

Rishi Bommasani, +113 more
- 16 Aug 2021 - 
TL;DR: The authors provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (e.g., language, vision, robotics, reasoning, human interaction) and technical principles(e. g.g. model architectures, training procedures, data, systems, security, evaluation, theory) to their applications.
Posted Content

A Framework for Understanding Sources of Harm throughout the Machine Learning Life Cycle

TL;DR: In this paper, the authors identify seven potential sources of downstream harm in machine learning, spanning data collection, development, and deployment, and propose a framework to facilitate more productive and precise communication around these issues.
Journal ArticleDOI

Good Proctor or "Big Brother"? Ethics of Online Exam Supervision Technologies.

TL;DR: In this paper, the authors provide a sustained moral philosophical analysis of online exam supervision technologies, focusing on ethical notions of academic integrity, fairness, non-maleficence, transparency, privacy, autonomy, liberty, and trust.
References
More filters
Journal ArticleDOI

An Integrative Model Of Organizational Trust

TL;DR: In this paper, a definition of trust and a model of its antecedents and outcomes are presented, which integrate research from multiple disciplines and differentiate trust from similar constructs, and several research propositions based on the model are presented.
Proceedings Article

Explaining and Harnessing Adversarial Examples

TL;DR: It is argued that the primary cause of neural networks' vulnerability to adversarial perturbation is their linear nature, supported by new quantitative results while giving the first explanation of the most intriguing fact about them: their generalization across architectures and training sets.
Proceedings Article

Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps

TL;DR: In this paper, the gradient of the class score with respect to the input image is computed to compute a class saliency map, which can be used for weakly supervised object segmentation using classification ConvNets.
Journal ArticleDOI

Trust as a Social Reality

TL;DR: Trust is seen to include both emotional and cognitive dimensions and to function as a deep assumption underwriting social order as mentioned in this paper, and trust is an underdeveloped concept in sociology, promising theoretical formulations are available in the recent work of Luhmann and Barber.
Journal ArticleDOI

Trust in Automation: Designing for Appropriate Reliance

TL;DR: This review considers trust from the organizational, sociological, interpersonal, psychological, and neurological perspectives, and considers how the context, automation characteristics, and cognitive processes affect the appropriateness of trust.
Related Papers (5)