This critical review explores two interrelated trends: the rapid increase in studies on machine learning (ML) applications within health informatics and the growing concerns about the reproducibility of these applications across different healthcare settings. Addressing these concerns necessitates acknowledging the uncertainty inherent in evaluating medical decision support systems. Therefore, we emphasize the importance of external validation and robustness assessment of the underlying ML models to better estimate their performance across diverse real-world scenarios. To raise awareness among health practitioners and ML researchers, we advocate for the widespread adoption of external validation practices and uncertainty quantification techniques. Our survey of specialized literature reveals that fewer than 4% of studies published in high-impact medical informatics journals over the past 13 years have validated their systems using data from settings different from those that provided the training data. This low percentage is incompatible with responsible research, given the potential risks posed by unreliable ML models in healthcare. Raising the standards for medical AI evaluation is crucial to improving practitioners' understanding of the potential and limitations of decision support systems in real-world settings. It is essential that uncertainty is not hidden in studies aimed at advancing knowledge in this field.

Marconi, L., Cabitza, F. (2025). Show and tell: A critical review on robustness and uncertainty for a more responsible medical AI. INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 202(October 2025) [10.1016/j.ijmedinf.2025.105970].

Show and tell: A critical review on robustness and uncertainty for a more responsible medical AI

Marconi, Luca
Co-primo
;
Cabitza, Federico
Co-primo
2025

Abstract

This critical review explores two interrelated trends: the rapid increase in studies on machine learning (ML) applications within health informatics and the growing concerns about the reproducibility of these applications across different healthcare settings. Addressing these concerns necessitates acknowledging the uncertainty inherent in evaluating medical decision support systems. Therefore, we emphasize the importance of external validation and robustness assessment of the underlying ML models to better estimate their performance across diverse real-world scenarios. To raise awareness among health practitioners and ML researchers, we advocate for the widespread adoption of external validation practices and uncertainty quantification techniques. Our survey of specialized literature reveals that fewer than 4% of studies published in high-impact medical informatics journals over the past 13 years have validated their systems using data from settings different from those that provided the training data. This low percentage is incompatible with responsible research, given the potential risks posed by unreliable ML models in healthcare. Raising the standards for medical AI evaluation is crucial to improving practitioners' understanding of the potential and limitations of decision support systems in real-world settings. It is essential that uncertainty is not hidden in studies aimed at advancing knowledge in this field.
Articolo in rivista - Review Essay
External validation; Machine learning; Medical informatics; Reproducibility; Robustness; Uncertainty evaluation; Uncertainty quantification
English
19-mag-2025
2025
202
October 2025
105970
open
Marconi, L., Cabitza, F. (2025). Show and tell: A critical review on robustness and uncertainty for a more responsible medical AI. INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 202(October 2025) [10.1016/j.ijmedinf.2025.105970].
File in questo prodotto:
File Dimensione Formato  
Marconi-Cabitza-2025-International Journal of Medical Informatics-VoR.pdf

accesso aperto

Descrizione: This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
Tipologia di allegato: Publisher’s Version (Version of Record, VoR)
Licenza: Creative Commons
Dimensione 1.31 MB
Formato Adobe PDF
1.31 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/557123
Citazioni
  • Scopus ND
  • ???jsp.display-item.citation.isi??? 0
Social impact