Unified Unsupervised Salient Object Detection via Knowledge Transfer
Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2024-07 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Yao, Yuan Liu, Wutao Gao, Pan Dai, Qun Qin, Jie |
description | Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3. |
doi_str_mv | 10.48550/arxiv.2404.14759 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2404_14759</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3044857587</sourcerecordid><originalsourceid>FETCH-LOGICAL-a959-9c00031541f473bb80f3ba0ae067ea9e36d3e5373da5f816fbea236956b0278e3</originalsourceid><addsrcrecordid>eNotkFFLwzAUhYMgOOZ-gE8WfG5Nc5OmfZRNnTjYg91zSdYbSalpTdqq_966-XQ4cDjnfpeQm5QmPBeC3iv_baeEccqTlEtRXJAFA0jjnDN2RVYhNJRSlkkmBCzI5uCssVhHBxfGHv1kw2zeVGvRDdFeN3gcog0Os9jORZNV0avrvlqs3zEqvXLBoL8ml0a1AVf_uiTl02O53sa7_fPL-mEXq0IUcXGchyEVPDVcgtY5NaAVVUgziapAyGpAARJqJUyeZkajYpAVItOUyRxhSW7PtSfEqvf2Q_mf6g-1OqHOibtzovfd54hhqJpu9G6-qQLK5_9IkUv4BYtSVik</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3044857587</pqid></control><display><type>article</type><title>Unified Unsupervised Salient Object Detection via Knowledge Transfer</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Yao, Yuan ; Liu, Wutao ; Gao, Pan ; Dai, Qun ; Qin, Jie</creator><creatorcontrib>Yao, Yuan ; Liu, Wutao ; Gao, Pan ; Dai, Qun ; Qin, Jie</creatorcontrib><description>Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2404.14759</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Annotations ; Computer Science - Computer Vision and Pattern Recognition ; Distillation ; Knowledge acquisition ; Knowledge management ; Labels ; Multimedia ; Object recognition ; Salience</subject><ispartof>arXiv.org, 2024-07</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2404.14759$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.24963/ijcai.2024/179$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Yao, Yuan</creatorcontrib><creatorcontrib>Liu, Wutao</creatorcontrib><creatorcontrib>Gao, Pan</creatorcontrib><creatorcontrib>Dai, Qun</creatorcontrib><creatorcontrib>Qin, Jie</creatorcontrib><title>Unified Unsupervised Salient Object Detection via Knowledge Transfer</title><title>arXiv.org</title><description>Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.</description><subject>Annotations</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Distillation</subject><subject>Knowledge acquisition</subject><subject>Knowledge management</subject><subject>Labels</subject><subject>Multimedia</subject><subject>Object recognition</subject><subject>Salience</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotkFFLwzAUhYMgOOZ-gE8WfG5Nc5OmfZRNnTjYg91zSdYbSalpTdqq_966-XQ4cDjnfpeQm5QmPBeC3iv_baeEccqTlEtRXJAFA0jjnDN2RVYhNJRSlkkmBCzI5uCssVhHBxfGHv1kw2zeVGvRDdFeN3gcog0Os9jORZNV0avrvlqs3zEqvXLBoL8ml0a1AVf_uiTl02O53sa7_fPL-mEXq0IUcXGchyEVPDVcgtY5NaAVVUgziapAyGpAARJqJUyeZkajYpAVItOUyRxhSW7PtSfEqvf2Q_mf6g-1OqHOibtzovfd54hhqJpu9G6-qQLK5_9IkUv4BYtSVik</recordid><startdate>20240713</startdate><enddate>20240713</enddate><creator>Yao, Yuan</creator><creator>Liu, Wutao</creator><creator>Gao, Pan</creator><creator>Dai, Qun</creator><creator>Qin, Jie</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240713</creationdate><title>Unified Unsupervised Salient Object Detection via Knowledge Transfer</title><author>Yao, Yuan ; Liu, Wutao ; Gao, Pan ; Dai, Qun ; Qin, Jie</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a959-9c00031541f473bb80f3ba0ae067ea9e36d3e5373da5f816fbea236956b0278e3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Annotations</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Distillation</topic><topic>Knowledge acquisition</topic><topic>Knowledge management</topic><topic>Labels</topic><topic>Multimedia</topic><topic>Object recognition</topic><topic>Salience</topic><toplevel>online_resources</toplevel><creatorcontrib>Yao, Yuan</creatorcontrib><creatorcontrib>Liu, Wutao</creatorcontrib><creatorcontrib>Gao, Pan</creatorcontrib><creatorcontrib>Dai, Qun</creatorcontrib><creatorcontrib>Qin, Jie</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yao, Yuan</au><au>Liu, Wutao</au><au>Gao, Pan</au><au>Dai, Qun</au><au>Qin, Jie</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Unified Unsupervised Salient Object Detection via Knowledge Transfer</atitle><jtitle>arXiv.org</jtitle><date>2024-07-13</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Recently, unsupervised salient object detection (USOD) has gained increasing attention due to its annotation-free nature. However, current methods mainly focus on specific tasks such as RGB and RGB-D, neglecting the potential for task migration. In this paper, we propose a unified USOD framework for generic USOD tasks. Firstly, we propose a Progressive Curriculum Learning-based Saliency Distilling (PCL-SD) mechanism to extract saliency cues from a pre-trained deep network. This mechanism starts with easy samples and progressively moves towards harder ones, to avoid initial interference caused by hard samples. Afterwards, the obtained saliency cues are utilized to train a saliency detector, and we employ a Self-rectify Pseudo-label Refinement (SPR) mechanism to improve the quality of pseudo-labels. Finally, an adapter-tuning method is devised to transfer the acquired saliency knowledge, leveraging shared knowledge to attain superior transferring performance on the target tasks. Extensive experiments on five representative SOD tasks confirm the effectiveness and feasibility of our proposed method. Code and supplement materials are available at https://github.com/I2-Multimedia-Lab/A2S-v3.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2404.14759</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-07 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_2404_14759 |
source | arXiv.org; Free E- Journals |
subjects | Annotations Computer Science - Computer Vision and Pattern Recognition Distillation Knowledge acquisition Knowledge management Labels Multimedia Object recognition Salience |
title | Unified Unsupervised Salient Object Detection via Knowledge Transfer |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-04T19%3A55%3A47IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Unified%20Unsupervised%20Salient%20Object%20Detection%20via%20Knowledge%20Transfer&rft.jtitle=arXiv.org&rft.au=Yao,%20Yuan&rft.date=2024-07-13&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2404.14759&rft_dat=%3Cproquest_arxiv%3E3044857587%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3044857587&rft_id=info:pmid/&rfr_iscdi=true |