Unified Unsupervised Salient Object Detection via Knowledge Transfer

Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-07
Hauptverfasser: Yao, Yuan, Liu, Wutao, Gao, Pan, Dai, Qun, Qin, Jie
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Yao, Yuan
Liu, Wutao
Gao, Pan
Dai, Qun
Qin, Jie
description Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.
doi_str_mv 10.48550/arxiv.2404.14759
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2404_14759</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3044857587</sourcerecordid><originalsourceid>FETCH-LOGICAL-a959-9c00031541f473bb80f3ba0ae067ea9e36d3e5373da5f816fbea236956b0278e3</originalsourceid><addsrcrecordid>eNotkFFLwzAUhYMgOOZ-gE8WfG5Nc5OmfZRNnTjYg91zSdYbSalpTdqq_966-XQ4cDjnfpeQm5QmPBeC3iv_baeEccqTlEtRXJAFA0jjnDN2RVYhNJRSlkkmBCzI5uCssVhHBxfGHv1kw2zeVGvRDdFeN3gcog0Os9jORZNV0avrvlqs3zEqvXLBoL8ml0a1AVf_uiTl02O53sa7_fPL-mEXq0IUcXGchyEVPDVcgtY5NaAVVUgziapAyGpAARJqJUyeZkajYpAVItOUyRxhSW7PtSfEqvf2Q_mf6g-1OqHOibtzovfd54hhqJpu9G6-qQLK5_9IkUv4BYtSVik</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3044857587</pqid></control><display><type>article</type><title>Unified Unsupervised Salient Object Detection via Knowledge Transfer</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Yao, Yuan ; Liu, Wutao ; Gao, Pan ; Dai, Qun ; Qin, Jie</creator><creatorcontrib>Yao, Yuan ; Liu, Wutao ; Gao, Pan ; Dai, Qun ; Qin, Jie</creatorcontrib><description>Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2404.14759</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Annotations ; Computer Science - Computer Vision and Pattern Recognition ; Distillation ; Knowledge acquisition ; Knowledge management ; Labels ; Multimedia ; Object recognition ; Salience</subject><ispartof>arXiv.org, 2024-07</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2404.14759$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.24963/ijcai.2024/179$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Yao, Yuan</creatorcontrib><creatorcontrib>Liu, Wutao</creatorcontrib><creatorcontrib>Gao, Pan</creatorcontrib><creatorcontrib>Dai, Qun</creatorcontrib><creatorcontrib>Qin, Jie</creatorcontrib><title>Unified Unsupervised Salient Object Detection via Knowledge Transfer</title><title>arXiv.org</title><description>Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.</description><subject>Annotations</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Distillation</subject><subject>Knowledge acquisition</subject><subject>Knowledge management</subject><subject>Labels</subject><subject>Multimedia</subject><subject>Object recognition</subject><subject>Salience</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotkFFLwzAUhYMgOOZ-gE8WfG5Nc5OmfZRNnTjYg91zSdYbSalpTdqq_966-XQ4cDjnfpeQm5QmPBeC3iv_baeEccqTlEtRXJAFA0jjnDN2RVYhNJRSlkkmBCzI5uCssVhHBxfGHv1kw2zeVGvRDdFeN3gcog0Os9jORZNV0avrvlqs3zEqvXLBoL8ml0a1AVf_uiTl02O53sa7_fPL-mEXq0IUcXGchyEVPDVcgtY5NaAVVUgziapAyGpAARJqJUyeZkajYpAVItOUyRxhSW7PtSfEqvf2Q_mf6g-1OqHOibtzovfd54hhqJpu9G6-qQLK5_9IkUv4BYtSVik</recordid><startdate>20240713</startdate><enddate>20240713</enddate><creator>Yao, Yuan</creator><creator>Liu, Wutao</creator><creator>Gao, Pan</creator><creator>Dai, Qun</creator><creator>Qin, Jie</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240713</creationdate><title>Unified Unsupervised Salient Object Detection via Knowledge Transfer</title><author>Yao, Yuan ; Liu, Wutao ; Gao, Pan ; Dai, Qun ; Qin, Jie</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a959-9c00031541f473bb80f3ba0ae067ea9e36d3e5373da5f816fbea236956b0278e3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Annotations</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Distillation</topic><topic>Knowledge acquisition</topic><topic>Knowledge management</topic><topic>Labels</topic><topic>Multimedia</topic><topic>Object recognition</topic><topic>Salience</topic><toplevel>online_resources</toplevel><creatorcontrib>Yao, Yuan</creatorcontrib><creatorcontrib>Liu, Wutao</creatorcontrib><creatorcontrib>Gao, Pan</creatorcontrib><creatorcontrib>Dai, Qun</creatorcontrib><creatorcontrib>Qin, Jie</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yao, Yuan</au><au>Liu, Wutao</au><au>Gao, Pan</au><au>Dai, Qun</au><au>Qin, Jie</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Unified Unsupervised Salient Object Detection via Knowledge Transfer</atitle><jtitle>arXiv.org</jtitle><date>2024-07-13</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2404.14759</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-07
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2404_14759
source arXiv.org; Free E- Journals
subjects Annotations
Computer Science - Computer Vision and Pattern Recognition
Distillation
Knowledge acquisition
Knowledge management
Labels
Multimedia
Object recognition
Salience
title Unified Unsupervised Salient Object Detection via Knowledge Transfer
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T19%3A55%3A47IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Unified%20Unsupervised%20Salient%20Object%20Detection%20via%20Knowledge%20Transfer&rft.jtitle=arXiv.org&rft.au=Yao,%20Yuan&rft.date=2024-07-13&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2404.14759&rft_dat=%3Cproquest_arxiv%3E3044857587%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3044857587&rft_id=info:pmid/&rfr_iscdi=true