Superpixel attention guided network for accurate and real-time salient object detection

Edge information has been proven to be effective for remedying the unclear boundaries of salient objects. Current salient object detection (SOD) methods usually utilize edge detection as an auxiliary task to introduce explicit edge information. However, edge detection is unable to provide the indisp...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Multimedia tools and applications 2022-11, Vol.81 (27), p.38921-38944
Hauptverfasser: Zhou, Zhiheng, Guo, Yongfan, Huang, Junchu, Dai, Ming, Deng, Ming, Yu, Qingjun
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Edge information has been proven to be effective for remedying the unclear boundaries of salient objects. Current salient object detection (SOD) methods usually utilize edge detection as an auxiliary task to introduce explicit edge information. However, edge detection is unable to provide the indispensable regional information for SOD, which may result in incomplete salient objects. To alleviate this risk, observing that superpixels hold the inherent property that contains both edge and regional information, we propose a superpixel attention guided network (SAGN) in this paper. Specifically, we first devise a novel supervised deep superpixel clustering (DSC) method to form the relation between superpixels and SOD. Based on the DSC, we build a superpixel attention module (SAM), which provides superpixel attention maps that can neatly separate different salient foreground and background regions, while preserving accurate boundaries of salient objects. Under the guidance of the SAM, a lightweight decoder with a simple but effective structure is able to yield high-quality salient objects with accurate and sharp boundaries. Hence, our model only contains less than 5 million parameters and achieves a real-time speed of around 40 FPS. Whilst offering a lightweight model and fast speed, our method still outperforms other 11 state-of-the-art approaches on six benchmark datasets.
ISSN:1380-7501
1573-7721
DOI:10.1007/s11042-022-13083-9