Automatic quality control of brain T1-weighted magnetic resonance images for a clinical data warehouse
•We propose a framework for the automatic QC of 3D T1w brain MRI for a clinical data warehouse.•We manually labeled 5500 images to train/test different convolutional neural networks.•The automatic approach can identify images which are not proper T1w brain MRIs (e.g. truncated images).•It is able to...
Gespeichert in:
Veröffentlicht in: | Medical image analysis 2022-01, Vol.75, p.102219-102219, Article 102219 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | •We propose a framework for the automatic QC of 3D T1w brain MRI for a clinical data warehouse.•We manually labeled 5500 images to train/test different convolutional neural networks.•The automatic approach can identify images which are not proper T1w brain MRIs (e.g. truncated images).•It is able to identify acquisitions for which gadolinium was injected.•It can also accurately identify low quality images.
[Display omitted]
Many studies on machine learning (ML) for computer-aided diagnosis have so far been mostly restricted to high-quality research data. Clinical data warehouses, gathering routine examinations from hospitals, offer great promises for training and validation of ML models in a realistic setting. However, the use of such clinical data warehouses requires quality control (QC) tools. Visual QC by experts is time-consuming and does not scale to large datasets. In this paper, we propose a convolutional neural network (CNN) for the automatic QC of 3D T1-weighted brain MRI for a large heterogeneous clinical data warehouse. To that purpose, we used the data warehouse of the hospitals of the Greater Paris area (Assistance Publique-Hôpitaux de Paris [AP-HP]). Specifically, the objectives were: 1) to identify images which are not proper T1-weighted brain MRIs; 2) to identify acquisitions for which gadolinium was injected; 3) to rate the overall image quality. We used 5000 images for training and validation and a separate set of 500 images for testing. In order to train/validate the CNN, the data were annotated by two trained raters according to a visual QC protocol that we specifically designed for application in the setting of a data warehouse. For objectives 1 and 2, our approach achieved excellent accuracy (balanced accuracy and F1-score >90%), similar to the human raters. For objective 3, the performance was good but substantially lower than that of human raters. Nevertheless, the automatic approach accurately identified (balanced accuracy and F1-score >80%) low quality images, which would typically need to be excluded. Overall, our approach shall be useful for exploiting hospital data warehouses in medical image computing. |
---|---|
ISSN: | 1361-8415 1361-8423 |
DOI: | 10.1016/j.media.2021.102219 |