Mammography classification with multi-view deep learning techniques: Investigating graph and transformer-based architectures
The potential and promise of deep learning systems to provide an independent assessment and relieve radiologists’ burden in screening mammography have been recognized in several studies. However, the low cancer prevalence, the need to process high-resolution images, and the need to combine informati...
Gespeichert in:
Veröffentlicht in: | Medical image analysis 2025-01, Vol.99, p.103320, Article 103320 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | The potential and promise of deep learning systems to provide an independent assessment and relieve radiologists’ burden in screening mammography have been recognized in several studies. However, the low cancer prevalence, the need to process high-resolution images, and the need to combine information from multiple views and scales still pose technical challenges. Multi-view architectures that combine information from the four mammographic views to produce an exam-level classification score are a promising approach to the automated processing of screening mammography. However, training such architectures from exam-level labels, without relying on pixel-level supervision, requires very large datasets and may result in suboptimal accuracy. Emerging architectures such as Visual Transformers (ViT) and graph-based architectures can potentially integrate ipsi-lateral and contra-lateral breast views better than traditional convolutional neural networks, thanks to their stronger ability of modeling long-range dependencies. In this paper, we extensively evaluate novel transformer-based and graph-based architectures against state-of-the-art multi-view convolutional neural networks, trained in a weakly-supervised setting on a middle-scale dataset, both in terms of performance and interpretability. Extensive experiments on the CSAW dataset suggest that, while transformer-based architecture outperform other architectures, different inductive biases lead to complementary strengths and weaknesses, as each architecture is sensitive to different signs and mammographic features. Hence, an ensemble of different architectures should be preferred over a winner-takes-all approach to achieve more accurate and robust results. Overall, the findings highlight the potential of a wide range of multi-view architectures for breast cancer classification, even in datasets of relatively modest size, although the detection of small lesions remains challenging without pixel-wise supervision or ad-hoc networks.
•Extended transformer-based architectures for handling four mammographic views.•Novel cross-view attention module for ipsi-lateral and contra-lateral analysis.•Experimentally compared transformers, convolutional and graph-based architectures.•Transformers with Swin-v2 backbone achieve highest individual performance.•Different architectures have complementary strengths and weaknesses. |
---|---|
ISSN: | 1361-8415 1361-8423 1361-8423 |
DOI: | 10.1016/j.media.2024.103320 |