MoMA: Momentum contrastive learning with multi-head attention-based knowledge distillation for histopathology image analysis
There is no doubt that advanced artificial intelligence models and high quality data are the keys to success in developing computational pathology tools. Although the overall volume of pathology data keeps increasing, a lack of quality data is a common issue when it comes to a specific task due to s...
Gespeichert in:
Veröffentlicht in: | Medical image analysis 2025-04, Vol.101, p.103421, Article 103421 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | There is no doubt that advanced artificial intelligence models and high quality data are the keys to success in developing computational pathology tools. Although the overall volume of pathology data keeps increasing, a lack of quality data is a common issue when it comes to a specific task due to several reasons including privacy and ethical issues with patient data. In this work, we propose to exploit knowledge distillation, i.e., utilize the existing model to learn a new, target model, to overcome such issues in computational pathology. Specifically, we employ a student–teacher framework to learn a target model from a pre-trained, teacher model without direct access to source data and distill relevant knowledge via momentum contrastive learning with multi-head attention mechanism, which provides consistent and context-aware feature representations. This enables the target model to assimilate informative representations of the teacher model while seamlessly adapting to the unique nuances of the target data. The proposed method is rigorously evaluated across different scenarios where the teacher model was trained on the same, relevant, and irrelevant classification tasks with the target model. Experimental results demonstrate the accuracy and robustness of our approach in transferring knowledge to different domains and tasks, outperforming other related methods. Moreover, the results provide a guideline on the learning strategy for different types of tasks and scenarios in computational pathology.
•MoMA is an efficient and effective learning framework for computational pathology.•MoMA improves knowledge distillation and transfer on a limited pathology dataset.•MoMA outperforms other related works in learning a target model for a specific task.•We investigate MoMA for same-, relevant-, and irrelevant-task distillation scenarios.•We provide a guideline on the learning strategy when limited datasets are available. |
---|---|
ISSN: | 1361-8415 1361-8423 1361-8423 |
DOI: | 10.1016/j.media.2024.103421 |