Multi-scale fully convolutional neural networks for histopathology image segmentation: From nuclear aberrations to the global tissue architecture

•Extensive integration of widely different spatial scales, as a “mimicry” of how humans approach analogous tasks, boosts the performance of a standard fully convolutional neural network in cancer segmentation in histopathology images, as shown on three different publicly available datasets.•A family...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Medical image analysis 2021-05, Vol.70, p.101996-101996, Article 101996
Hauptverfasser: Schmitz, Rüdiger, Madesta, Frederic, Nielsen, Maximilian, Krause, Jenny, Steurer, Stefan, Werner, René, Rösch, Thomas
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•Extensive integration of widely different spatial scales, as a “mimicry” of how humans approach analogous tasks, boosts the performance of a standard fully convolutional neural network in cancer segmentation in histopathology images, as shown on three different publicly available datasets.•A family of U-Net-based architectures as human operator-inspired multi-scale multi-encoder networks is proposed. The approach can be easily adopted to any encoder-decoder segmentation architecture and extended to multiple path fusions.•By use of an additional classification loss, additional encoders for largely different spatial scales as the target scale can be trained in a memory-efficient fashion and with moderate additional cost. [Display omitted] Histopathologic diagnosis relies on simultaneous integration of information from a broad range of scales, ranging from nuclear aberrations (≈O(0.1μm)) through cellular structures (≈O(10μm)) to the global tissue architecture (⪆O(1mm)). To explicitly mimic how human pathologists combine multi-scale information, we introduce a family of multi-encoder fully-convolutional neural networks with deep fusion. We present a simple block for merging model paths with differing spatial scales in a spatial relationship-preserving fashion, which can readily be included in standard encoder-decoder networks. Additionally, a context classification gate block is proposed as an alternative for the incorporation of global context. Our experiments were performed on three publicly available whole-slide images of recent challenges (PAIP 2019: hepatocellular carcinoma segmentation; BACH 2020: breast cancer segmentation; CAMELYON 2016: metastasis detection in lymph nodes). The multi-scale architectures consistently outperformed the baseline single-scale U-Nets by a large margin. They benefit from local as well as global context and particularly a combination of both. If feature maps from different scales are fused, doing so in a manner preserving spatial relationships was found to be beneficial. Deep guidance by a context classification loss appeared to improve model training at low computational costs. All multi-scale models had a reduced GPU memory footprint compared to ensembles of individual U-Nets trained on different image scales. Additional path fusions were shown to be possible at low computational cost, opening up possibilities for further, systematic and task-specific architecture optimisation. The findings demonstrate the potential of the
ISSN:1361-8415
1361-8423
DOI:10.1016/j.media.2021.101996