Deep learning segmentation model for quantification of infarct size in pigs with myocardial ischemia/reperfusion

Infarct size (IS) is the most robust end point for evaluating the success of preclinical studies on cardioprotection. The gold standard for IS quantification in ischemia/reperfusion (I/R) experiments is triphenyl tetrazolium chloride (TTC) staining, typically done manually. This study aimed to deter...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Basic research in cardiology 2024-12, Vol.119 (6), p.923-936
Hauptverfasser: Braczko, Felix, Skyschally, Andreas, Lieder, Helmut, Kather, Jakob Nikolas, Kleinbongard, Petra, Heusch, Gerd
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Infarct size (IS) is the most robust end point for evaluating the success of preclinical studies on cardioprotection. The gold standard for IS quantification in ischemia/reperfusion (I/R) experiments is triphenyl tetrazolium chloride (TTC) staining, typically done manually. This study aimed to determine if automation through deep learning segmentation is a time-saving and valid alternative to standard IS quantification. High-resolution images from TTC-stained, macroscopic heart slices were retrospectively collected from pig experiments ( n  = 390) with I/R without/with cardioprotection to cover a wide IS range. Existing IS data from pig experiments, quantified using a standard method of manual and subsequent digital labeling of film-scan annotations, were used as reference. To automate the evaluation process with the aim to be more objective and save time, a deep learning pipeline was implemented; the collected images ( n  = 3869) were pre-processed by cropping and labeled (image annotations). To ensure their usability as training data for a deep learning segmentation model, IS was quantified from image annotations and compared to IS quantified using the existing film-scan annotations. A supervised deep learning segmentation model based on dynamic U-Net architecture was developed and trained. The evaluation of the trained model was performed by fivefold cross-validation ( n  = 220 experiments) and testing on an independent test set ( n  = 170 experiments). Performance metrics (Dice similarity coefficient [DSC], pixel accuracy [ACC], average precision [mAP]) were calculated. IS was then quantified from predictions and compared to IS quantified from image annotations (linear regression, Pearson’s r ; analysis of covariance; Bland–Altman plots). Performance metrics near 1 indicated a strong model performance on cross-validated data (DSC: 0.90, ACC: 0.98, mAP: 0.90) and on the test set data (DSC: 0.89, ACC: 0.98, mAP: 0.93). IS quantified from predictions correlated well with IS quantified from image annotations in all data sets (cross-validation: r  = 0.98; test data set: r  = 0.95) and analysis of covariance identified no significant differences. The model reduced the IS quantification time per experiment from approximately 90 min to 20 s. The model was further tested on a preliminary test set from experiments in isolated, saline-perfused rat hearts with regional I/R without/with cardioprotection ( n  = 27). There was also no significant difference in IS bet
ISSN:1435-1803
0300-8428
1435-1803
DOI:10.1007/s00395-024-01081-x