Iterative multi-path tracking for video and volume segmentation with sparse point supervision
highlights•Ground-truth annotation are necessary to train machine learning models.•We annotate video and volumetric sequences using a single 2D point per frame.•No constraints on appearance, shape, and motion/displacement of object of interest.•Promising results on surgical tool and slitlamp videos,...
Gespeichert in:
Veröffentlicht in: | Medical image analysis 2018-12, Vol.50, p.65-81 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | highlights•Ground-truth annotation are necessary to train machine learning models.•We annotate video and volumetric sequences using a single 2D point per frame.•No constraints on appearance, shape, and motion/displacement of object of interest.•Promising results on surgical tool and slitlamp videos, brain MRI, CT scans of inner ear.
[Display omitted]
Recent machine learning strategies for segmentation tasks have shown great ability when trained on large pixel-wise annotated image datasets. It remains a major challenge however to aggregate such datasets, as the time and monetary cost associated with collecting extensive annotations is extremely high. This is particularly the case for generating precise pixel-wise annotations in video and volumetric image data. To this end, this work presents a novel framework to produce pixel-wise segmentations using minimal supervision. Our method relies on 2D point supervision, whereby a single 2D location within an object of interest is provided on each image of the data. Our method then estimates the object appearance in a semi-supervised fashion by learning object-image-specific features and by using these in a semi-supervised learning framework. Our object model is then used in a graph-based optimization problem that takes into account all provided locations and the image data in order to infer the complete pixel-wise segmentation. In practice, we solve this optimally as a tracking problem using a K-shortest path approach. Both the object model and segmentation are then refined iteratively to further improve the final segmentation. We show that by collecting 2D locations using a gaze tracker, our approach can provide state-of-the-art segmentations on a range of objects and image modalities (video and 3D volumes), and that these can then be used to train supervised machine learning classifiers. |
---|---|
ISSN: | 1361-8415 1361-8423 |
DOI: | 10.1016/j.media.2018.08.007 |