DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query
Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object de...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2023-12 |
---|---|
Hauptverfasser: | , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Zhang, Yuhang Deng, Yuang Zhang, Xiaopeng Li, Jie Qiu, Robert C Tian, Qi |
description | Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection. |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2907600185</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2907600185</sourcerecordid><originalsourceid>FETCH-proquest_journals_29076001853</originalsourceid><addsrcrecordid>eNqNissKgkAYRocgSMp3GGgtjGNeahdZtHCT1FoG_bURmbG5FPX0DdIDtPo43zkz5NEoCoNsQ-kC-Vr3hBCapDSOIw_dcijKHd7Xhj8BF8CU4KLDrVQ4BwPulgK_uLk7rKUdB2hwIWs28A-bHBMNLp3qBJ_4YkG9V2jeskGD_9slWp-O18M5GJV8WNCm6qVVwqmKbkmaEBJmcfRf9QVloEBd</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2907600185</pqid></control><display><type>article</type><title>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</title><source>Free E- Journals</source><creator>Zhang, Yuhang ; Deng, Yuang ; Zhang, Xiaopeng ; Li, Jie ; Qiu, Robert C ; Tian, Qi</creator><creatorcontrib>Zhang, Yuhang ; Deng, Yuang ; Zhang, Xiaopeng ; Li, Jie ; Qiu, Robert C ; Tian, Qi</creatorcontrib><description>Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Active learning ; Annotations ; Budgets ; Decoupling ; Labeling ; Labels ; Learning ; Localization ; Object recognition ; Queries</subject><ispartof>arXiv.org, 2023-12</ispartof><rights>2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Zhang, Yuhang</creatorcontrib><creatorcontrib>Deng, Yuang</creatorcontrib><creatorcontrib>Zhang, Xiaopeng</creatorcontrib><creatorcontrib>Li, Jie</creatorcontrib><creatorcontrib>Qiu, Robert C</creatorcontrib><creatorcontrib>Tian, Qi</creatorcontrib><title>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</title><title>arXiv.org</title><description>Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.</description><subject>Active learning</subject><subject>Annotations</subject><subject>Budgets</subject><subject>Decoupling</subject><subject>Labeling</subject><subject>Labels</subject><subject>Learning</subject><subject>Localization</subject><subject>Object recognition</subject><subject>Queries</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNissKgkAYRocgSMp3GGgtjGNeahdZtHCT1FoG_bURmbG5FPX0DdIDtPo43zkz5NEoCoNsQ-kC-Vr3hBCapDSOIw_dcijKHd7Xhj8BF8CU4KLDrVQ4BwPulgK_uLk7rKUdB2hwIWs28A-bHBMNLp3qBJ_4YkG9V2jeskGD_9slWp-O18M5GJV8WNCm6qVVwqmKbkmaEBJmcfRf9QVloEBd</recordid><startdate>20231228</startdate><enddate>20231228</enddate><creator>Zhang, Yuhang</creator><creator>Deng, Yuang</creator><creator>Zhang, Xiaopeng</creator><creator>Li, Jie</creator><creator>Qiu, Robert C</creator><creator>Tian, Qi</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20231228</creationdate><title>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</title><author>Zhang, Yuhang ; Deng, Yuang ; Zhang, Xiaopeng ; Li, Jie ; Qiu, Robert C ; Tian, Qi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_29076001853</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Active learning</topic><topic>Annotations</topic><topic>Budgets</topic><topic>Decoupling</topic><topic>Labeling</topic><topic>Labels</topic><topic>Learning</topic><topic>Localization</topic><topic>Object recognition</topic><topic>Queries</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Yuhang</creatorcontrib><creatorcontrib>Deng, Yuang</creatorcontrib><creatorcontrib>Zhang, Xiaopeng</creatorcontrib><creatorcontrib>Li, Jie</creatorcontrib><creatorcontrib>Qiu, Robert C</creatorcontrib><creatorcontrib>Tian, Qi</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Yuhang</au><au>Deng, Yuang</au><au>Zhang, Xiaopeng</au><au>Li, Jie</au><au>Qiu, Robert C</au><au>Tian, Qi</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query</atitle><jtitle>arXiv.org</jtitle><date>2023-12-28</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Active learning has been demonstrated effective to reduce labeling cost, while most progress has been designed for image recognition, there still lacks instance-level active learning for object detection. In this paper, we rethink two key components, i.e., localization and recognition, for object detection, and find that the correctness of them are highly related, therefore, it is not necessary to annotate both boxes and classes if we are given pseudo annotations provided with the trained model. Motivated by this, we propose an efficient query strategy, termed as DeLR, that Decoupling the Localization and Recognition for active query. In this way, we are probably free of class annotations when the localization is correct, and able to assign the labeling budget for more informative samples. There are two main differences in DeLR: 1) Unlike previous methods mostly focus on image-level annotations, where the queried samples are selected and exhausted annotated. In DeLR, the query is based on region-level, and we only annotate the object region that is queried; 2) Instead of directly providing both localization and recognition annotations, we separately query the two components, and thus reduce the recognition budget with the pseudo class labels provided by the model. Experiments on several benchmarks demonstrate its superiority. We hope our proposed query strategy would shed light on researches in active learning in object detection.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2023-12 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2907600185 |
source | Free E- Journals |
subjects | Active learning Annotations Budgets Decoupling Labeling Labels Learning Localization Object recognition Queries |
title | DeLR: Active Learning for Detection with Decoupled Localization and Recognition Query |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T10%3A56%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=DeLR:%20Active%20Learning%20for%20Detection%20with%20Decoupled%20Localization%20and%20Recognition%20Query&rft.jtitle=arXiv.org&rft.au=Zhang,%20Yuhang&rft.date=2023-12-28&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2907600185%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2907600185&rft_id=info:pmid/&rfr_iscdi=true |