Artificial Intelligence (ai) systems are precious support for decision-making, with many applications also in the medical domain. The interaction between mds and ai enjoys a renewed interest following the increased possibilities of deep learning devices. However, we still have limited evidence-based knowledge of the context, design, and psychological mechanisms that craft an optimal human–ai collaboration. In this multicentric study, 21 endoscopists reviewed 504 videos of lesions prospectively acquired from real colonoscopies. They were asked to provide an optical diagnosis with and without the assistance of an ai support system. Endoscopists were influenced by ai (OR=3.05), but not erratically: they followed the ai advice more when it was correct (OR=3.48) than incorrect (OR=1.85). Endoscopists achieved this outcome through a weighted integration of their and the ai opinions, considering the case-by-case estimations of the two reliabilities. This Bayesian-like rational behavior allowed the human–ai hybrid team to outperform both agents taken alone. We discuss the features of the human–ai interaction that determined this favorable outcome.

Reverberi, C., Rigon, T., Solari, A., Hassan, C., Cherubini, P., Antonelli, G., et al. (2022). Experimental evidence of effective human–AI collaboration in medical decision-making. SCIENTIFIC REPORTS, 12(1) [10.1038/s41598-022-18751-2].

Experimental evidence of effective human–AI collaboration in medical decision-making

Reverberi C.
Primo
;
Rigon T.;Solari A.;Cherubini P.;
2022

Abstract

Artificial Intelligence (ai) systems are precious support for decision-making, with many applications also in the medical domain. The interaction between mds and ai enjoys a renewed interest following the increased possibilities of deep learning devices. However, we still have limited evidence-based knowledge of the context, design, and psychological mechanisms that craft an optimal human–ai collaboration. In this multicentric study, 21 endoscopists reviewed 504 videos of lesions prospectively acquired from real colonoscopies. They were asked to provide an optical diagnosis with and without the assistance of an ai support system. Endoscopists were influenced by ai (OR=3.05), but not erratically: they followed the ai advice more when it was correct (OR=3.48) than incorrect (OR=1.85). Endoscopists achieved this outcome through a weighted integration of their and the ai opinions, considering the case-by-case estimations of the two reliabilities. This Bayesian-like rational behavior allowed the human–ai hybrid team to outperform both agents taken alone. We discuss the features of the human–ai interaction that determined this favorable outcome.
Articolo in rivista - Articolo scientifico
Bayes Theorem; Humans; Artificial Intelligence; Clinical Decision-Making
English
2-set-2022
2022
12
1
14952
open
Reverberi, C., Rigon, T., Solari, A., Hassan, C., Cherubini, P., Antonelli, G., et al. (2022). Experimental evidence of effective human–AI collaboration in medical decision-making. SCIENTIFIC REPORTS, 12(1) [10.1038/s41598-022-18751-2].
File in questo prodotto:
File Dimensione Formato  
10281-397813_VoR.pdf

accesso aperto

Tipologia di allegato: Publisher’s Version (Version of Record, VoR)
Licenza: Creative Commons
Dimensione 1.29 MB
Formato Adobe PDF
1.29 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/397813
Citazioni
  • Scopus 34
  • ???jsp.display-item.citation.isi??? 31
Social impact