Federated Foundation Model for Cardiac CT Imaging
Federated learning (FL) is a renowned technique for utilizing decentralized data while preserving privacy. However, real-world applications often involve inherent challenges such as partially labeled datasets, where not all clients possess expert annotations of all labels of interest, leaving large...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , , , , , , , , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Federated learning (FL) is a renowned technique for utilizing decentralized
data while preserving privacy. However, real-world applications often involve
inherent challenges such as partially labeled datasets, where not all clients
possess expert annotations of all labels of interest, leaving large portions of
unlabeled data unused. In this study, we conduct the largest federated cardiac
CT imaging analysis to date, focusing on partially labeled datasets ($n=8,124$)
of Transcatheter Aortic Valve Implantation (TAVI) patients over eight hospital
clients. Transformer architectures, which are the major building blocks of
current foundation models, have shown superior performance when trained on
larger cohorts than traditional CNNs. However, when trained on small
task-specific labeled sample sizes, it is currently not feasible to exploit
their underlying attention mechanism for improved performance. Therefore, we
developed a two-stage semi-supervised learning strategy that distills knowledge
from several task-specific CNNs (landmark detection and segmentation of
calcification) into a single transformer model by utilizing large amounts of
unlabeled data typically residing unused in hospitals to mitigate these issues.
This method not only improves the predictive accuracy and generalizability of
transformer-based architectures but also facilitates the simultaneous learning
of all partial labels within a single transformer model across the federation.
Additionally, we show that our transformer-based model extracts more meaningful
features for further downstream tasks than the UNet-based one by only training
the last layer to also solve segmentation of coronary arteries. We make the
code and weights of the final model openly available, which can serve as a
foundation model for further research in cardiac CT imaging. |
---|---|
DOI: | 10.48550/arxiv.2407.07557 |