Limits of trust in medical AI

Journal of Medical Ethics:medethics-2019-105935 (forthcoming)
Download Edit this record How to cite View on PhilPapers
Abstract
Artificial intelligence is expected to revolutionise the practice of medicine. Recent advancements in the field of deep learning have demonstrated success in variety of clinical tasks: detecting diabetic retinopathy from images, predicting hospital readmissions, aiding in the discovery of new drugs, etc. AI’s progress in medicine, however, has led to concerns regarding the potential effects of this technology on relationships of trust in clinical practice. In this paper, I will argue that there is merit to these concerns, since AI systems can be relied on, and are capable of reliability, but cannot be trusted, and are not capable of trustworthiness. Insofar as patients are required to rely on AI systems for their medical decision-making, there is potential for this to produce a deficit of trust in relationships in clinical practice.
Keywords
No keywords specified (fix it)
Categories
(categorize this paper)
PhilPapers/Archive ID
HATLOT
Revision history
Archival date: 2020-05-17
View upload history
References found in this work BETA

No references found.

Add more references

Citations of this work BETA

No citations found.

Add more citations

Added to PP index
2020-03-28

Total views
19 ( #48,228 of 49,046 )

Recent downloads (6 months)
19 ( #32,443 of 49,046 )

How can I increase my downloads?

Downloads since first upload
This graph includes both downloads from PhilArchive and clicks to external links.