Active Learning with Expected Error Reduction
Active learning has been studied extensively as a method for efficient data collection. Among the many approaches in literature, Expected Error Reduction (EER) (Roy and McCallum) has been shown to be an effective method for active learning: select the candidate sample that, in expectation, maximally...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Mussmann, Stephen Reisler, Julia Tsai, Daniel Mousavi, Ehsan O'Brien, Shayne Goldszmidt, Moises |
description | Active learning has been studied extensively as a method for efficient data
collection. Among the many approaches in literature, Expected Error Reduction
(EER) (Roy and McCallum) has been shown to be an effective method for active
learning: select the candidate sample that, in expectation, maximally decreases
the error on an unlabeled set. However, EER requires the model to be retrained
for every candidate sample and thus has not been widely used for modern deep
neural networks due to this large computational cost. In this paper we
reformulate EER under the lens of Bayesian active learning and derive a
computationally efficient version that can use any Bayesian parameter sampling
method (such as arXiv:1506.02142). We then compare the empirical performance of
our method using Monte Carlo dropout for parameter sampling against state of
the art methods in the deep active learning literature. Experiments are
performed on four standard benchmark datasets and three WILDS datasets
(arXiv:2012.07421). The results indicate that our method outperforms all other
methods except one in the data shift scenario: a model dependent,
non-information theoretic method that requires an order of magnitude higher
computational cost (arXiv:1906.03671). |
doi_str_mv | 10.48550/arxiv.2211.09283 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2211_09283</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2211_09283</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-77f32e1be092504bd7fbfd47ad576c6b4c63d9ed4a40227377b88205c751cc273</originalsourceid><addsrcrecordid>eNotjsuKwkAURHvjYlA_YFb2DyTTz9y4DBIdISCI-9CPG23QRNqMj783o64KiuLUIeSbs1TlWrMfE-_hmgrBecrmIpdfJClcH65IKzSxDe2e3kJ_oOX9jK5HT8sYu0i36P-GWddOyKgxxwtOPzkmu2W5W_wm1Wa1XhRVYjKQCUAjBXKLw4dmynpobOMVGK8hc5lVLpN-jl4ZxYQACWDzXDDtQHPnhmJMZm_sy7c-x3Ay8VH_e9cvb_kEKvk8Uw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Active Learning with Expected Error Reduction</title><source>arXiv.org</source><creator>Mussmann, Stephen ; Reisler, Julia ; Tsai, Daniel ; Mousavi, Ehsan ; O'Brien, Shayne ; Goldszmidt, Moises</creator><creatorcontrib>Mussmann, Stephen ; Reisler, Julia ; Tsai, Daniel ; Mousavi, Ehsan ; O'Brien, Shayne ; Goldszmidt, Moises</creatorcontrib><description>Active learning has been studied extensively as a method for efficient data
collection. Among the many approaches in literature, Expected Error Reduction
(EER) (Roy and McCallum) has been shown to be an effective method for active
learning: select the candidate sample that, in expectation, maximally decreases
the error on an unlabeled set. However, EER requires the model to be retrained
for every candidate sample and thus has not been widely used for modern deep
neural networks due to this large computational cost. In this paper we
reformulate EER under the lens of Bayesian active learning and derive a
computationally efficient version that can use any Bayesian parameter sampling
method (such as arXiv:1506.02142). We then compare the empirical performance of
our method using Monte Carlo dropout for parameter sampling against state of
the art methods in the deep active learning literature. Experiments are
performed on four standard benchmark datasets and three WILDS datasets
(arXiv:2012.07421). The results indicate that our method outperforms all other
methods except one in the data shift scenario: a model dependent,
non-information theoretic method that requires an order of magnitude higher
computational cost (arXiv:1906.03671).</description><identifier>DOI: 10.48550/arxiv.2211.09283</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2022-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2211.09283$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2211.09283$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Mussmann, Stephen</creatorcontrib><creatorcontrib>Reisler, Julia</creatorcontrib><creatorcontrib>Tsai, Daniel</creatorcontrib><creatorcontrib>Mousavi, Ehsan</creatorcontrib><creatorcontrib>O'Brien, Shayne</creatorcontrib><creatorcontrib>Goldszmidt, Moises</creatorcontrib><title>Active Learning with Expected Error Reduction</title><description>Active learning has been studied extensively as a method for efficient data
collection. Among the many approaches in literature, Expected Error Reduction
(EER) (Roy and McCallum) has been shown to be an effective method for active
learning: select the candidate sample that, in expectation, maximally decreases
the error on an unlabeled set. However, EER requires the model to be retrained
for every candidate sample and thus has not been widely used for modern deep
neural networks due to this large computational cost. In this paper we
reformulate EER under the lens of Bayesian active learning and derive a
computationally efficient version that can use any Bayesian parameter sampling
method (such as arXiv:1506.02142). We then compare the empirical performance of
our method using Monte Carlo dropout for parameter sampling against state of
the art methods in the deep active learning literature. Experiments are
performed on four standard benchmark datasets and three WILDS datasets
(arXiv:2012.07421). The results indicate that our method outperforms all other
methods except one in the data shift scenario: a model dependent,
non-information theoretic method that requires an order of magnitude higher
computational cost (arXiv:1906.03671).</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjsuKwkAURHvjYlA_YFb2DyTTz9y4DBIdISCI-9CPG23QRNqMj783o64KiuLUIeSbs1TlWrMfE-_hmgrBecrmIpdfJClcH65IKzSxDe2e3kJ_oOX9jK5HT8sYu0i36P-GWddOyKgxxwtOPzkmu2W5W_wm1Wa1XhRVYjKQCUAjBXKLw4dmynpobOMVGK8hc5lVLpN-jl4ZxYQACWDzXDDtQHPnhmJMZm_sy7c-x3Ay8VH_e9cvb_kEKvk8Uw</recordid><startdate>20221116</startdate><enddate>20221116</enddate><creator>Mussmann, Stephen</creator><creator>Reisler, Julia</creator><creator>Tsai, Daniel</creator><creator>Mousavi, Ehsan</creator><creator>O'Brien, Shayne</creator><creator>Goldszmidt, Moises</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221116</creationdate><title>Active Learning with Expected Error Reduction</title><author>Mussmann, Stephen ; Reisler, Julia ; Tsai, Daniel ; Mousavi, Ehsan ; O'Brien, Shayne ; Goldszmidt, Moises</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-77f32e1be092504bd7fbfd47ad576c6b4c63d9ed4a40227377b88205c751cc273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Mussmann, Stephen</creatorcontrib><creatorcontrib>Reisler, Julia</creatorcontrib><creatorcontrib>Tsai, Daniel</creatorcontrib><creatorcontrib>Mousavi, Ehsan</creatorcontrib><creatorcontrib>O'Brien, Shayne</creatorcontrib><creatorcontrib>Goldszmidt, Moises</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Mussmann, Stephen</au><au>Reisler, Julia</au><au>Tsai, Daniel</au><au>Mousavi, Ehsan</au><au>O'Brien, Shayne</au><au>Goldszmidt, Moises</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Active Learning with Expected Error Reduction</atitle><date>2022-11-16</date><risdate>2022</risdate><abstract>Active learning has been studied extensively as a method for efficient data
collection. Among the many approaches in literature, Expected Error Reduction
(EER) (Roy and McCallum) has been shown to be an effective method for active
learning: select the candidate sample that, in expectation, maximally decreases
the error on an unlabeled set. However, EER requires the model to be retrained
for every candidate sample and thus has not been widely used for modern deep
neural networks due to this large computational cost. In this paper we
reformulate EER under the lens of Bayesian active learning and derive a
computationally efficient version that can use any Bayesian parameter sampling
method (such as arXiv:1506.02142). We then compare the empirical performance of
our method using Monte Carlo dropout for parameter sampling against state of
the art methods in the deep active learning literature. Experiments are
performed on four standard benchmark datasets and three WILDS datasets
(arXiv:2012.07421). The results indicate that our method outperforms all other
methods except one in the data shift scenario: a model dependent,
non-information theoretic method that requires an order of magnitude higher
computational cost (arXiv:1906.03671).</abstract><doi>10.48550/arxiv.2211.09283</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2211.09283 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2211_09283 |
source | arXiv.org |
subjects | Computer Science - Learning |
title | Active Learning with Expected Error Reduction |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-19T09%3A05%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Active%20Learning%20with%20Expected%20Error%20Reduction&rft.au=Mussmann,%20Stephen&rft.date=2022-11-16&rft_id=info:doi/10.48550/arxiv.2211.09283&rft_dat=%3Carxiv_GOX%3E2211_09283%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |