Prediction of in-plane organ deformation during free-breathing radiotherapy via discriminative spatial transformer networks

•Free-breathing motion compensation aimed at image-guided radiotherapy.•Novel framework with multi-scale feature extraction and spatial transformation for in-plane motion prediction and future frame generation.•Prediction of the next k deformations from a given input image sequence.•Validation on di...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Medical image analysis 2020-08, Vol.64, p.101754-101754, Article 101754
Hauptverfasser: Romaguera, Liset Vázquez, Plantefève, Rosalie, Romero, Francisco Perdigón, Hébert, François, Carrier, Jean-François, Kadoury, Samuel
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•Free-breathing motion compensation aimed at image-guided radiotherapy.•Novel framework with multi-scale feature extraction and spatial transformation for in-plane motion prediction and future frame generation.•Prediction of the next k deformations from a given input image sequence.•Validation on different imaging modalities: MRI, US and CT. [Display omitted] External beam radiotherapy is a commonly used treatment option for patients with cancer in the thoracic and abdominal regions. However, respiratory motion constitutes a major limitation during the intervention. It may stray the pre-defined target and trajectories determined during planning from the actual anatomy. We propose a novel framework to predict the in-plane organ motion. We introduce a recurrent encoder-decoder architecture which leverages feature representations at multiple scales. It simultaneously learns to map dense deformations between consecutive images from a given input sequence and to extrapolate them through time. Subsequently, several cascade-arranged spatial transformers use the predicted deformation fields to generate a future image sequence. We propose the use of a composite loss function which minimizes the difference between ground-truth and predicted images while maintaining smooth deformations. Our model is trained end-to-end in an unsupervised manner, thus it does not require additional information beyond image data. Moreover, no pre-processing steps such as segmentation or registration are needed. We report results on 85 different cases (healthy subjects and patients) belonging to multiples datasets across different imaging modalities. Experiments were aimed at investigating the importance of the proposed multi-scale architecture design and the effect of increasing the number of predicted frames on the overall accuracy of the model. The proposed model was able to predict vessel positions in the next temporal image with a median accuracy of 0.45 (0.55) mm, 0.45 (0.74) mm and 0.28 (0.58) mm in MRI, US and CT datasets, respectively. The obtained results show the strong potential of the model by achieving accurate matching between the predicted and target images on several imaging modalities.
ISSN:1361-8415
1361-8423
DOI:10.1016/j.media.2020.101754