Active Learning with Expected Error Reduction

Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Mussmann, Stephen, Reisler, Julia, Tsai, Daniel, Mousavi, Ehsan, O'Brien, Shayne, Goldszmidt, Moises
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Mussmann, Stephen
Reisler, Julia
Tsai, Daniel
Mousavi, Ehsan
O'Brien, Shayne
Goldszmidt, Moises
description Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally decreases the error on an unlabeled set. However, EER requires the model to be retrained for every candidate sample and thus has not been widely used for modern deep neural networks due to this large computational cost. In this paper we reformulate EER under the lens of Bayesian active learning and derive a computationally efficient version that can use any Bayesian parameter sampling method (such as arXiv:1506.02142). We then compare the empirical performance of our method using Monte Carlo dropout for parameter sampling against state of the art methods in the deep active learning literature. Experiments are performed on four standard benchmark datasets and three WILDS datasets (arXiv:2012.07421). The results indicate that our method outperforms all other methods except one in the data shift scenario: a model dependent, non-information theoretic method that requires an order of magnitude higher computational cost (arXiv:1906.03671).
doi_str_mv 10.48550/arxiv.2211.09283
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2211_09283</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2211_09283</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-77f32e1be092504bd7fbfd47ad576c6b4c63d9ed4a40227377b88205c751cc273</originalsourceid><addsrcrecordid>eNotjsuKwkAURHvjYlA_YFb2DyTTz9y4DBIdISCI-9CPG23QRNqMj783o64KiuLUIeSbs1TlWrMfE-_hmgrBecrmIpdfJClcH65IKzSxDe2e3kJ_oOX9jK5HT8sYu0i36P-GWddOyKgxxwtOPzkmu2W5W_wm1Wa1XhRVYjKQCUAjBXKLw4dmynpobOMVGK8hc5lVLpN-jl4ZxYQACWDzXDDtQHPnhmJMZm_sy7c-x3Ay8VH_e9cvb_kEKvk8Uw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Active Learning with Expected Error Reduction</title><source>arXiv.org</source><creator>Mussmann, Stephen ; Reisler, Julia ; Tsai, Daniel ; Mousavi, Ehsan ; O'Brien, Shayne ; Goldszmidt, Moises</creator><creatorcontrib>Mussmann, Stephen ; Reisler, Julia ; Tsai, Daniel ; Mousavi, Ehsan ; O'Brien, Shayne ; Goldszmidt, Moises</creatorcontrib><description>Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally decreases the error on an unlabeled set. However, EER requires the model to be retrained for every candidate sample and thus has not been widely used for modern deep neural networks due to this large computational cost. In this paper we reformulate EER under the lens of Bayesian active learning and derive a computationally efficient version that can use any Bayesian parameter sampling method (such as arXiv:1506.02142). We then compare the empirical performance of our method using Monte Carlo dropout for parameter sampling against state of the art methods in the deep active learning literature. Experiments are performed on four standard benchmark datasets and three WILDS datasets (arXiv:2012.07421). The results indicate that our method outperforms all other methods except one in the data shift scenario: a model dependent, non-information theoretic method that requires an order of magnitude higher computational cost (arXiv:1906.03671).</description><identifier>DOI: 10.48550/arxiv.2211.09283</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2022-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2211.09283$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2211.09283$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Mussmann, Stephen</creatorcontrib><creatorcontrib>Reisler, Julia</creatorcontrib><creatorcontrib>Tsai, Daniel</creatorcontrib><creatorcontrib>Mousavi, Ehsan</creatorcontrib><creatorcontrib>O'Brien, Shayne</creatorcontrib><creatorcontrib>Goldszmidt, Moises</creatorcontrib><title>Active Learning with Expected Error Reduction</title><description>Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally decreases the error on an unlabeled set. However, EER requires the model to be retrained for every candidate sample and thus has not been widely used for modern deep neural networks due to this large computational cost. In this paper we reformulate EER under the lens of Bayesian active learning and derive a computationally efficient version that can use any Bayesian parameter sampling method (such as arXiv:1506.02142). We then compare the empirical performance of our method using Monte Carlo dropout for parameter sampling against state of the art methods in the deep active learning literature. Experiments are performed on four standard benchmark datasets and three WILDS datasets (arXiv:2012.07421). The results indicate that our method outperforms all other methods except one in the data shift scenario: a model dependent, non-information theoretic method that requires an order of magnitude higher computational cost (arXiv:1906.03671).</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjsuKwkAURHvjYlA_YFb2DyTTz9y4DBIdISCI-9CPG23QRNqMj783o64KiuLUIeSbs1TlWrMfE-_hmgrBecrmIpdfJClcH65IKzSxDe2e3kJ_oOX9jK5HT8sYu0i36P-GWddOyKgxxwtOPzkmu2W5W_wm1Wa1XhRVYjKQCUAjBXKLw4dmynpobOMVGK8hc5lVLpN-jl4ZxYQACWDzXDDtQHPnhmJMZm_sy7c-x3Ay8VH_e9cvb_kEKvk8Uw</recordid><startdate>20221116</startdate><enddate>20221116</enddate><creator>Mussmann, Stephen</creator><creator>Reisler, Julia</creator><creator>Tsai, Daniel</creator><creator>Mousavi, Ehsan</creator><creator>O'Brien, Shayne</creator><creator>Goldszmidt, Moises</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221116</creationdate><title>Active Learning with Expected Error Reduction</title><author>Mussmann, Stephen ; Reisler, Julia ; Tsai, Daniel ; Mousavi, Ehsan ; O'Brien, Shayne ; Goldszmidt, Moises</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-77f32e1be092504bd7fbfd47ad576c6b4c63d9ed4a40227377b88205c751cc273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Mussmann, Stephen</creatorcontrib><creatorcontrib>Reisler, Julia</creatorcontrib><creatorcontrib>Tsai, Daniel</creatorcontrib><creatorcontrib>Mousavi, Ehsan</creatorcontrib><creatorcontrib>O'Brien, Shayne</creatorcontrib><creatorcontrib>Goldszmidt, Moises</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Mussmann, Stephen</au><au>Reisler, Julia</au><au>Tsai, Daniel</au><au>Mousavi, Ehsan</au><au>O'Brien, Shayne</au><au>Goldszmidt, Moises</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Active Learning with Expected Error Reduction</atitle><date>2022-11-16</date><risdate>2022</risdate><abstract>Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally decreases the error on an unlabeled set. However, EER requires the model to be retrained for every candidate sample and thus has not been widely used for modern deep neural networks due to this large computational cost. In this paper we reformulate EER under the lens of Bayesian active learning and derive a computationally efficient version that can use any Bayesian parameter sampling method (such as arXiv:1506.02142). We then compare the empirical performance of our method using Monte Carlo dropout for parameter sampling against state of the art methods in the deep active learning literature. Experiments are performed on four standard benchmark datasets and three WILDS datasets (arXiv:2012.07421). The results indicate that our method outperforms all other methods except one in the data shift scenario: a model dependent, non-information theoretic method that requires an order of magnitude higher computational cost (arXiv:1906.03671).</abstract><doi>10.48550/arxiv.2211.09283</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2211.09283
ispartof
issn
language eng
recordid cdi_arxiv_primary_2211_09283
source arXiv.org
subjects Computer Science - Learning
title Active Learning with Expected Error Reduction
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-19T09%3A05%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Active%20Learning%20with%20Expected%20Error%20Reduction&rft.au=Mussmann,%20Stephen&rft.date=2022-11-16&rft_id=info:doi/10.48550/arxiv.2211.09283&rft_dat=%3Carxiv_GOX%3E2211_09283%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true