Robust and data-efficient generalization of self-supervised machine learning for diagnostic imaging

Machine-learning models for medical tasks can match or surpass the performance of clinical experts. However, in settings differing from those of the training dataset, the performance of a model can deteriorate substantially. Here we report a representation-learning strategy for machine-learning mode...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Nature biomedical engineering 2023-06, Vol.7 (6), p.756-779
Hauptverfasser: Azizi, Shekoofeh, Culp, Laura, Freyberg, Jan, Mustafa, Basil, Baur, Sebastien, Kornblith, Simon, Chen, Ting, Tomasev, Nenad, Mitrović, Jovana, Strachan, Patricia, Mahdavi, S. Sara, Wulczyn, Ellery, Babenko, Boris, Walker, Megan, Loh, Aaron, Chen, Po-Hsuan Cameron, Liu, Yuan, Bavishi, Pinal, McKinney, Scott Mayer, Winkens, Jim, Roy, Abhijit Guha, Beaver, Zach, Ryan, Fiona, Krogue, Justin, Etemadi, Mozziyar, Telang, Umesh, Liu, Yun, Peng, Lily, Corrado, Greg S., Webster, Dale R., Fleet, David, Hinton, Geoffrey, Houlsby, Neil, Karthikesalingam, Alan, Norouzi, Mohammad, Natarajan, Vivek
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Machine-learning models for medical tasks can match or surpass the performance of clinical experts. However, in settings differing from those of the training dataset, the performance of a model can deteriorate substantially. Here we report a representation-learning strategy for machine-learning models applied to medical-imaging tasks that mitigates such ‘out of distribution’ performance problem and that improves model robustness and training efficiency. The strategy, which we named REMEDIS (for ‘Robust and Efficient Medical Imaging with Self-supervision’), combines large-scale supervised transfer learning on natural images and intermediate contrastive self-supervised learning on medical images and requires minimal task-specific customization. We show the utility of REMEDIS in a range of diagnostic-imaging tasks covering six imaging domains and 15 test datasets, and by simulating three realistic out-of-distribution scenarios. REMEDIS improved in-distribution diagnostic accuracies up to 11.5% with respect to strong supervised baseline models, and in out-of-distribution settings required only 1–33% of the data for retraining to match the performance of supervised models retrained using all available data. REMEDIS may accelerate the development lifecycle of machine-learning models for medical imaging. A representation-learning strategy for machine-learning models applied to medical-imaging tasks improves model robustness and training efficiency and mitigates suboptimal out-of-distribution performance.
ISSN:2157-846X
2157-846X
DOI:10.1038/s41551-023-01049-7