z-logo
open-access-imgOpen Access
‘If You’re Going to Trust the Machine, Then That Trust Has Got to Be Based on Something’:
Author(s) -
Peter Winter,
Annamaria Carusi
Publication year - 2022
Publication title -
science and technology studies
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.675
H-Index - 14
ISSN - 2243-4690
DOI - 10.23987/sts.102198
Subject(s) - interpretability , negotiation , interrogation , computer science , artificial intelligence , knowledge management , psychology , data science , sociology , political science , social science , law
The role of Artificial Intelligence (AI) in clinical decision-making raises issues of trust. One issue concerns the conditions of trusting the AI which tends to be based on validation. However, little attention has been given to how validation is formed, how comparisons come to be accepted, and how AI algorithms are trusted in decision-making. Drawing on interviews with collaborative researchers developing three AI technologies for the early diagnosis of pulmonary hypertension (PH), we show how validation of the AI is jointly produced so that trust in the algorithm is built up through the negotiation of criteria and terms of comparison during interactions. These processes build up interpretability and interrogation, and co-constitute trust in the technology. As they do so, it becomes difficult to sustain a strict distinction between artificial and human/social intelligence.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here