PDZSeg: Adapting the Foundation Model for Dissection Zone Segmentation with Visual Prompts in Robot-assisted Endoscopic Submucosal Dissection
Purpose: Endoscopic surgical environments present challenges for dissection zone segmentation due to unclear boundaries between tissue types, leading to segmentation errors where models misidentify or overlook edges. This study aims to provide precise dissection zone suggestions during endoscopic su...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Purpose: Endoscopic surgical environments present challenges for dissection
zone segmentation due to unclear boundaries between tissue types, leading to
segmentation errors where models misidentify or overlook edges. This study aims
to provide precise dissection zone suggestions during endoscopic submucosal
dissection (ESD) procedures, enhancing ESD safety.
Methods: We propose the Prompted-based Dissection Zone Segmentation (PDZSeg)
model, designed to leverage diverse visual prompts such as scribbles and
bounding boxes. By overlaying these prompts onto images and fine-tuning a
foundational model on a specialized dataset, our approach improves segmentation
performance and user experience through flexible input methods.
Results: The PDZSeg model was validated using three experimental setups:
in-domain evaluation, variability in visual prompt availability, and robustness
assessment. Using the ESD-DZSeg dataset, results show that our method
outperforms state-of-the-art segmentation approaches. This is the first study
to integrate visual prompt design into dissection zone segmentation.
Conclusion: The PDZSeg model effectively utilizes visual prompts to enhance
segmentation performance and user experience, supported by the novel ESD-DZSeg
dataset as a benchmark for dissection zone segmentation in ESD. Our work
establishes a foundation for future research. |
---|---|
DOI: | 10.48550/arxiv.2411.18169 |