DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query

Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object de...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2023-12
Hauptverfasser: Zhang, Yuhang, Deng, Yuang, Zhang, Xiaopeng, Li, Jie, Qiu, Robert C, Tian, Qi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Zhang, Yuhang
Deng, Yuang
Zhang, Xiaopeng
Li, Jie
Qiu, Robert C
Tian, Qi
description Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2907600185</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2907600185</sourcerecordid><originalsourceid>FETCH-proquest_journals_29076001853</originalsourceid><addsrcrecordid>eNqNissKgkAYRocgSMp3GGgtjGNeahdZtHCT1FoG_bURmbG5FPX0DdIDtPo43zkz5NEoCoNsQ-kC-Vr3hBCapDSOIw_dcijKHd7Xhj8BF8CU4KLDrVQ4BwPulgK_uLk7rKUdB2hwIWs28A-bHBMNLp3qBJ_4YkG9V2jeskGD_9slWp-O18M5GJV8WNCm6qVVwqmKbkmaEBJmcfRf9QVloEBd</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2907600185</pqid></control><display><type>article</type><title>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</title><source>Free E- Journals</source><creator>Zhang, Yuhang ; Deng, Yuang ; Zhang, Xiaopeng ; Li, Jie ; Qiu, Robert C ; Tian, Qi</creator><creatorcontrib>Zhang, Yuhang ; Deng, Yuang ; Zhang, Xiaopeng ; Li, Jie ; Qiu, Robert C ; Tian, Qi</creatorcontrib><description>Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Active learning ; Annotations ; Budgets ; Decoupling ; Labeling ; Labels ; Learning ; Localization ; Object recognition ; Queries</subject><ispartof>arXiv.org, 2023-12</ispartof><rights>2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Zhang, Yuhang</creatorcontrib><creatorcontrib>Deng, Yuang</creatorcontrib><creatorcontrib>Zhang, Xiaopeng</creatorcontrib><creatorcontrib>Li, Jie</creatorcontrib><creatorcontrib>Qiu, Robert C</creatorcontrib><creatorcontrib>Tian, Qi</creatorcontrib><title>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</title><title>arXiv.org</title><description>Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.</description><subject>Active learning</subject><subject>Annotations</subject><subject>Budgets</subject><subject>Decoupling</subject><subject>Labeling</subject><subject>Labels</subject><subject>Learning</subject><subject>Localization</subject><subject>Object recognition</subject><subject>Queries</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNissKgkAYRocgSMp3GGgtjGNeahdZtHCT1FoG_bURmbG5FPX0DdIDtPo43zkz5NEoCoNsQ-kC-Vr3hBCapDSOIw_dcijKHd7Xhj8BF8CU4KLDrVQ4BwPulgK_uLk7rKUdB2hwIWs28A-bHBMNLp3qBJ_4YkG9V2jeskGD_9slWp-O18M5GJV8WNCm6qVVwqmKbkmaEBJmcfRf9QVloEBd</recordid><startdate>20231228</startdate><enddate>20231228</enddate><creator>Zhang, Yuhang</creator><creator>Deng, Yuang</creator><creator>Zhang, Xiaopeng</creator><creator>Li, Jie</creator><creator>Qiu, Robert C</creator><creator>Tian, Qi</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20231228</creationdate><title>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</title><author>Zhang, Yuhang ; Deng, Yuang ; Zhang, Xiaopeng ; Li, Jie ; Qiu, Robert C ; Tian, Qi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_29076001853</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Active learning</topic><topic>Annotations</topic><topic>Budgets</topic><topic>Decoupling</topic><topic>Labeling</topic><topic>Labels</topic><topic>Learning</topic><topic>Localization</topic><topic>Object recognition</topic><topic>Queries</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Yuhang</creatorcontrib><creatorcontrib>Deng, Yuang</creatorcontrib><creatorcontrib>Zhang, Xiaopeng</creatorcontrib><creatorcontrib>Li, Jie</creatorcontrib><creatorcontrib>Qiu, Robert C</creatorcontrib><creatorcontrib>Tian, Qi</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Yuhang</au><au>Deng, Yuang</au><au>Zhang, Xiaopeng</au><au>Li, Jie</au><au>Qiu, Robert C</au><au>Tian, Qi</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</atitle><jtitle>arXiv.org</jtitle><date>2023-12-28</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-12
issn 2331-8422
language eng
recordid cdi_proquest_journals_2907600185
source Free E- Journals
subjects Active learning
Annotations
Budgets
Decoupling
Labeling
Labels
Learning
Localization
Object recognition
Queries
title DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T10%3A56%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=DeLR:%20Active%20Learning%20for%20Detection%20with%20Decoupled%20Localization%20and%20Recognition%20Query&rft.jtitle=arXiv.org&rft.au=Zhang,%20Yuhang&rft.date=2023-12-28&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2907600185%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2907600185&rft_id=info:pmid/&rfr_iscdi=true