A Formal Account of AI Trustworthiness: Connecting Intrinsic and Perceived Trustworthiness

Aies '24: Proceedings of the 2024 Aaai/Acmconference on Ai, Ethics, and Society (forthcoming)
  Copy   BIBTEX

Abstract

This paper proposes a formal account of AI trustworthiness, connecting both intrinsic and perceived trustworthiness in an operational schematization. We argue that trustworthiness extends beyond the inherent capabilities of an AI system to include significant influences from observers' perceptions, such as perceived transparency, agency locus, and human oversight. While the concept of perceived trustworthiness is discussed in the literature, few attempts have been made to connect it with the intrinsic trustworthiness of AI systems. Our analysis introduces a novel schematization to quantify trustworthiness by assessing the discrepancies between expected and observed behaviors and how these affect perceived uncertainty and trust. The paper provides a formalization for measuring trustworthiness, taking into account both perceived and intrinsic characteristics. By detailing the factors that influence trust, this study aims to foster more ethical and widely accepted AI technologies, ensuring they meet both functional and ethical criteria.

Author's Profile

Piercosma Bisconti
Scuola Superiore di Studi Universitari e di Perfezionamento Sant'Anna

Analytics

Added to PP
yesterday

Downloads
0

6 months
0

Historical graph of downloads since first upload
This graph includes both downloads from PhilArchive and clicks on external links on PhilPapers.
How can I increase my downloads?