A Survey of Deep Active Learning

Active learning (AL) attempts to maximize a model’s performance gain while annotating the fewest samples possible. Deep learning (DL) is greedy for data and requires a large amount of data supply to optimize a massive number of parameters if the model is to learn how to extract high-quality features...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:ACM computing surveys 2022-12, Vol.54 (9), p.1-40
Hauptverfasser: Ren, Pengzhen, Xiao, Yun, Chang, Xiaojun, Huang, Po-Yao, Li, Zhihui, Gupta, Brij B., Chen, Xiaojiang, Wang, Xin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 40
container_issue 9
container_start_page 1
container_title ACM computing surveys
container_volume 54
creator Ren, Pengzhen
Xiao, Yun
Chang, Xiaojun
Huang, Po-Yao
Li, Zhihui
Gupta, Brij B.
Chen, Xiaojiang
Wang, Xin
description Active learning (AL) attempts to maximize a model’s performance gain while annotating the fewest samples possible. Deep learning (DL) is greedy for data and requires a large amount of data supply to optimize a massive number of parameters if the model is to learn how to extract high-quality features. In recent years, due to the rapid development of internet technology, we have entered an era of information abundance characterized by massive amounts of available data. As a result, DL has attracted significant attention from researchers and has been rapidly developed. Compared with DL, however, researchers have a relatively low interest in AL. This is mainly because before the rise of DL, traditional machine learning requires relatively few labeled samples, meaning that early AL is rarely according the value it deserves. Although DL has made breakthroughs in various fields, most of this success is due to a large number of publicly available annotated datasets. However, the acquisition of a large number of high-quality annotated datasets consumes a lot of manpower, making it unfeasible in fields that require high levels of expertise (such as speech recognition, information extraction, medical images, etc.). Therefore, AL is gradually coming to receive the attention it is due. It is therefore natural to investigate whether AL can be used to reduce the cost of sample annotation while retaining the powerful learning capabilities of DL. As a result of such investigations, deep active learning (DeepAL) has emerged. Although research on this topic is quite abundant, there has not yet been a comprehensive survey of DeepAL-related works; accordingly, this article aims to fill this gap. We provide a formal classification method for the existing work, along with a comprehensive and systematic overview. In addition, we also analyze and summarize the development of DeepAL from an application perspective. Finally, we discuss the confusion and problems associated with DeepAL and provide some possible development directions.
doi_str_mv 10.1145/3472291
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2733415851</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2733415851</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-f4765386db1a2c3eefd17268a718358d80de89e2abff8e19d52fa82ac9af41ac3</originalsourceid><addsrcrecordid>eNotj81KxDAYRYMoWEfxFQIuXFXz5adJlmX8hYILdR0y6RfpoG1N2oF5eyszq7s53MMh5BrYHYBU90Jqzi2ckAKU0qUWEk5JwUTFSiYYOycXOW8ZY1xCVRBa0_c57XBPh0gfEEdah6nbIW3Qp77rvy7JWfTfGa-OuyKfT48f65eyeXt-XddNGQTYqYxSV0qYqt2A50EgxhY0r4zXYIQyrWEtGovcb2I0CLZVPHrDfbA-SvBBrMjN4XdMw--MeXLbYU79onRciyVCGQULdXugQhpyThjdmLofn_YOmPvPd8d88Qdc6UmF</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2733415851</pqid></control><display><type>article</type><title>A Survey of Deep Active Learning</title><source>ACM Digital Library</source><creator>Ren, Pengzhen ; Xiao, Yun ; Chang, Xiaojun ; Huang, Po-Yao ; Li, Zhihui ; Gupta, Brij B. ; Chen, Xiaojiang ; Wang, Xin</creator><creatorcontrib>Ren, Pengzhen ; Xiao, Yun ; Chang, Xiaojun ; Huang, Po-Yao ; Li, Zhihui ; Gupta, Brij B. ; Chen, Xiaojiang ; Wang, Xin</creatorcontrib><description>Active learning (AL) attempts to maximize a model’s performance gain while annotating the fewest samples possible. Deep learning (DL) is greedy for data and requires a large amount of data supply to optimize a massive number of parameters if the model is to learn how to extract high-quality features. In recent years, due to the rapid development of internet technology, we have entered an era of information abundance characterized by massive amounts of available data. As a result, DL has attracted significant attention from researchers and has been rapidly developed. Compared with DL, however, researchers have a relatively low interest in AL. This is mainly because before the rise of DL, traditional machine learning requires relatively few labeled samples, meaning that early AL is rarely according the value it deserves. Although DL has made breakthroughs in various fields, most of this success is due to a large number of publicly available annotated datasets. However, the acquisition of a large number of high-quality annotated datasets consumes a lot of manpower, making it unfeasible in fields that require high levels of expertise (such as speech recognition, information extraction, medical images, etc.). Therefore, AL is gradually coming to receive the attention it is due. It is therefore natural to investigate whether AL can be used to reduce the cost of sample annotation while retaining the powerful learning capabilities of DL. As a result of such investigations, deep active learning (DeepAL) has emerged. Although research on this topic is quite abundant, there has not yet been a comprehensive survey of DeepAL-related works; accordingly, this article aims to fill this gap. We provide a formal classification method for the existing work, along with a comprehensive and systematic overview. In addition, we also analyze and summarize the development of DeepAL from an application perspective. Finally, we discuss the confusion and problems associated with DeepAL and provide some possible development directions.</description><identifier>ISSN: 0360-0300</identifier><identifier>EISSN: 1557-7341</identifier><identifier>DOI: 10.1145/3472291</identifier><language>eng</language><publisher>Baltimore: Association for Computing Machinery</publisher><subject>Active learning ; Annotations ; Computer science ; Datasets ; Deep learning ; Feature extraction ; Information retrieval ; Machine learning ; Medical imaging ; Object recognition ; Speech recognition</subject><ispartof>ACM computing surveys, 2022-12, Vol.54 (9), p.1-40</ispartof><rights>Copyright Association for Computing Machinery Dec 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-f4765386db1a2c3eefd17268a718358d80de89e2abff8e19d52fa82ac9af41ac3</citedby><cites>FETCH-LOGICAL-c319t-f4765386db1a2c3eefd17268a718358d80de89e2abff8e19d52fa82ac9af41ac3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27903,27904</link.rule.ids></links><search><creatorcontrib>Ren, Pengzhen</creatorcontrib><creatorcontrib>Xiao, Yun</creatorcontrib><creatorcontrib>Chang, Xiaojun</creatorcontrib><creatorcontrib>Huang, Po-Yao</creatorcontrib><creatorcontrib>Li, Zhihui</creatorcontrib><creatorcontrib>Gupta, Brij B.</creatorcontrib><creatorcontrib>Chen, Xiaojiang</creatorcontrib><creatorcontrib>Wang, Xin</creatorcontrib><title>A Survey of Deep Active Learning</title><title>ACM computing surveys</title><description>Active learning (AL) attempts to maximize a model’s performance gain while annotating the fewest samples possible. Deep learning (DL) is greedy for data and requires a large amount of data supply to optimize a massive number of parameters if the model is to learn how to extract high-quality features. In recent years, due to the rapid development of internet technology, we have entered an era of information abundance characterized by massive amounts of available data. As a result, DL has attracted significant attention from researchers and has been rapidly developed. Compared with DL, however, researchers have a relatively low interest in AL. This is mainly because before the rise of DL, traditional machine learning requires relatively few labeled samples, meaning that early AL is rarely according the value it deserves. Although DL has made breakthroughs in various fields, most of this success is due to a large number of publicly available annotated datasets. However, the acquisition of a large number of high-quality annotated datasets consumes a lot of manpower, making it unfeasible in fields that require high levels of expertise (such as speech recognition, information extraction, medical images, etc.). Therefore, AL is gradually coming to receive the attention it is due. It is therefore natural to investigate whether AL can be used to reduce the cost of sample annotation while retaining the powerful learning capabilities of DL. As a result of such investigations, deep active learning (DeepAL) has emerged. Although research on this topic is quite abundant, there has not yet been a comprehensive survey of DeepAL-related works; accordingly, this article aims to fill this gap. We provide a formal classification method for the existing work, along with a comprehensive and systematic overview. In addition, we also analyze and summarize the development of DeepAL from an application perspective. Finally, we discuss the confusion and problems associated with DeepAL and provide some possible development directions.</description><subject>Active learning</subject><subject>Annotations</subject><subject>Computer science</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Feature extraction</subject><subject>Information retrieval</subject><subject>Machine learning</subject><subject>Medical imaging</subject><subject>Object recognition</subject><subject>Speech recognition</subject><issn>0360-0300</issn><issn>1557-7341</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNotj81KxDAYRYMoWEfxFQIuXFXz5adJlmX8hYILdR0y6RfpoG1N2oF5eyszq7s53MMh5BrYHYBU90Jqzi2ckAKU0qUWEk5JwUTFSiYYOycXOW8ZY1xCVRBa0_c57XBPh0gfEEdah6nbIW3Qp77rvy7JWfTfGa-OuyKfT48f65eyeXt-XddNGQTYqYxSV0qYqt2A50EgxhY0r4zXYIQyrWEtGovcb2I0CLZVPHrDfbA-SvBBrMjN4XdMw--MeXLbYU79onRciyVCGQULdXugQhpyThjdmLofn_YOmPvPd8d88Qdc6UmF</recordid><startdate>20221201</startdate><enddate>20221201</enddate><creator>Ren, Pengzhen</creator><creator>Xiao, Yun</creator><creator>Chang, Xiaojun</creator><creator>Huang, Po-Yao</creator><creator>Li, Zhihui</creator><creator>Gupta, Brij B.</creator><creator>Chen, Xiaojiang</creator><creator>Wang, Xin</creator><general>Association for Computing Machinery</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20221201</creationdate><title>A Survey of Deep Active Learning</title><author>Ren, Pengzhen ; Xiao, Yun ; Chang, Xiaojun ; Huang, Po-Yao ; Li, Zhihui ; Gupta, Brij B. ; Chen, Xiaojiang ; Wang, Xin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-f4765386db1a2c3eefd17268a718358d80de89e2abff8e19d52fa82ac9af41ac3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Active learning</topic><topic>Annotations</topic><topic>Computer science</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Feature extraction</topic><topic>Information retrieval</topic><topic>Machine learning</topic><topic>Medical imaging</topic><topic>Object recognition</topic><topic>Speech recognition</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Ren, Pengzhen</creatorcontrib><creatorcontrib>Xiao, Yun</creatorcontrib><creatorcontrib>Chang, Xiaojun</creatorcontrib><creatorcontrib>Huang, Po-Yao</creatorcontrib><creatorcontrib>Li, Zhihui</creatorcontrib><creatorcontrib>Gupta, Brij B.</creatorcontrib><creatorcontrib>Chen, Xiaojiang</creatorcontrib><creatorcontrib>Wang, Xin</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>ACM computing surveys</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ren, Pengzhen</au><au>Xiao, Yun</au><au>Chang, Xiaojun</au><au>Huang, Po-Yao</au><au>Li, Zhihui</au><au>Gupta, Brij B.</au><au>Chen, Xiaojiang</au><au>Wang, Xin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Survey of Deep Active Learning</atitle><jtitle>ACM computing surveys</jtitle><date>2022-12-01</date><risdate>2022</risdate><volume>54</volume><issue>9</issue><spage>1</spage><epage>40</epage><pages>1-40</pages><issn>0360-0300</issn><eissn>1557-7341</eissn><abstract>Active learning (AL) attempts to maximize a model’s performance gain while annotating the fewest samples possible. Deep learning (DL) is greedy for data and requires a large amount of data supply to optimize a massive number of parameters if the model is to learn how to extract high-quality features. In recent years, due to the rapid development of internet technology, we have entered an era of information abundance characterized by massive amounts of available data. As a result, DL has attracted significant attention from researchers and has been rapidly developed. Compared with DL, however, researchers have a relatively low interest in AL. This is mainly because before the rise of DL, traditional machine learning requires relatively few labeled samples, meaning that early AL is rarely according the value it deserves. Although DL has made breakthroughs in various fields, most of this success is due to a large number of publicly available annotated datasets. However, the acquisition of a large number of high-quality annotated datasets consumes a lot of manpower, making it unfeasible in fields that require high levels of expertise (such as speech recognition, information extraction, medical images, etc.). Therefore, AL is gradually coming to receive the attention it is due. It is therefore natural to investigate whether AL can be used to reduce the cost of sample annotation while retaining the powerful learning capabilities of DL. As a result of such investigations, deep active learning (DeepAL) has emerged. Although research on this topic is quite abundant, there has not yet been a comprehensive survey of DeepAL-related works; accordingly, this article aims to fill this gap. We provide a formal classification method for the existing work, along with a comprehensive and systematic overview. In addition, we also analyze and summarize the development of DeepAL from an application perspective. Finally, we discuss the confusion and problems associated with DeepAL and provide some possible development directions.</abstract><cop>Baltimore</cop><pub>Association for Computing Machinery</pub><doi>10.1145/3472291</doi><tpages>40</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0360-0300
ispartof ACM computing surveys, 2022-12, Vol.54 (9), p.1-40
issn 0360-0300
1557-7341
language eng
recordid cdi_proquest_journals_2733415851
source ACM Digital Library
subjects Active learning
Annotations
Computer science
Datasets
Deep learning
Feature extraction
Information retrieval
Machine learning
Medical imaging
Object recognition
Speech recognition
title A Survey of Deep Active Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-23T05%3A07%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Survey%20of%20Deep%20Active%20Learning&rft.jtitle=ACM%20computing%20surveys&rft.au=Ren,%20Pengzhen&rft.date=2022-12-01&rft.volume=54&rft.issue=9&rft.spage=1&rft.epage=40&rft.pages=1-40&rft.issn=0360-0300&rft.eissn=1557-7341&rft_id=info:doi/10.1145/3472291&rft_dat=%3Cproquest_cross%3E2733415851%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2733415851&rft_id=info:pmid/&rfr_iscdi=true