Latent Outlier Exposure for Anomaly Detection with Contaminated Data

Proceedings of the 39th International Conference on Machine Learning, 2022, volume:162, pages:18153--18167 Anomaly detection aims at identifying data points that show systematic deviations from the majority of data in an unlabeled dataset. A common assumption is that clean training data (free of ano...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Qiu, Chen, Li, Aodong, Kloft, Marius, Rudolph, Maja, Mandt, Stephan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Qiu, Chen
Li, Aodong
Kloft, Marius
Rudolph, Maja
Mandt, Stephan
description Proceedings of the 39th International Conference on Machine Learning, 2022, volume:162, pages:18153--18167 Anomaly detection aims at identifying data points that show systematic deviations from the majority of data in an unlabeled dataset. A common assumption is that clean training data (free of anomalies) is available, which is often violated in practice. We propose a strategy for training an anomaly detector in the presence of unlabeled anomalies that is compatible with a broad class of models. The idea is to jointly infer binary labels to each datum (normal vs. anomalous) while updating the model parameters. Inspired by outlier exposure (Hendrycks et al., 2018) that considers synthetically created, labeled anomalies, we thereby use a combination of two losses that share parameters: one for the normal and one for the anomalous data. We then iteratively proceed with block coordinate updates on the parameters and the most likely (latent) labels. Our experiments with several backbone models on three image datasets, 30 tabular data sets, and a video anomaly detection benchmark showed consistent and significant improvements over the baselines.
doi_str_mv 10.48550/arxiv.2202.08088
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2202_08088</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2202_08088</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-e047abceb000baa974a7943cffc8112875a1bb7be8afc2bf0b4302bc8bec4443</originalsourceid><addsrcrecordid>eNotz7FOwzAUhWEvDKjwAEz4BRJuHAe7Y5WUghSpQ9mje821sJTEletC-_aFwnSm_0ifEA8VlNo2DTxhOoWvUilQJViw9lZ0PWaes9we8xg4yfVpHw_HxNLHJFdznHA8y44zuxziLL9D_pRtnDNOYf4pP2SHGe_EjcfxwPf_uxC7l_V7-1r0281bu-oLfDa2YNAGyTEBACEujUaz1LXz3tmqUtY0WBEZYoveKfJAugZFzhI7rXW9EI9_r1fFsE9hwnQefjXDVVNfAGD-RgA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Latent Outlier Exposure for Anomaly Detection with Contaminated Data</title><source>arXiv.org</source><creator>Qiu, Chen ; Li, Aodong ; Kloft, Marius ; Rudolph, Maja ; Mandt, Stephan</creator><creatorcontrib>Qiu, Chen ; Li, Aodong ; Kloft, Marius ; Rudolph, Maja ; Mandt, Stephan</creatorcontrib><description>Proceedings of the 39th International Conference on Machine Learning, 2022, volume:162, pages:18153--18167 Anomaly detection aims at identifying data points that show systematic deviations from the majority of data in an unlabeled dataset. A common assumption is that clean training data (free of anomalies) is available, which is often violated in practice. We propose a strategy for training an anomaly detector in the presence of unlabeled anomalies that is compatible with a broad class of models. The idea is to jointly infer binary labels to each datum (normal vs. anomalous) while updating the model parameters. Inspired by outlier exposure (Hendrycks et al., 2018) that considers synthetically created, labeled anomalies, we thereby use a combination of two losses that share parameters: one for the normal and one for the anomalous data. We then iteratively proceed with block coordinate updates on the parameters and the most likely (latent) labels. Our experiments with several backbone models on three image datasets, 30 tabular data sets, and a video anomaly detection benchmark showed consistent and significant improvements over the baselines.</description><identifier>DOI: 10.48550/arxiv.2202.08088</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2022-02</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2202.08088$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2202.08088$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Qiu, Chen</creatorcontrib><creatorcontrib>Li, Aodong</creatorcontrib><creatorcontrib>Kloft, Marius</creatorcontrib><creatorcontrib>Rudolph, Maja</creatorcontrib><creatorcontrib>Mandt, Stephan</creatorcontrib><title>Latent Outlier Exposure for Anomaly Detection with Contaminated Data</title><description>Proceedings of the 39th International Conference on Machine Learning, 2022, volume:162, pages:18153--18167 Anomaly detection aims at identifying data points that show systematic deviations from the majority of data in an unlabeled dataset. A common assumption is that clean training data (free of anomalies) is available, which is often violated in practice. We propose a strategy for training an anomaly detector in the presence of unlabeled anomalies that is compatible with a broad class of models. The idea is to jointly infer binary labels to each datum (normal vs. anomalous) while updating the model parameters. Inspired by outlier exposure (Hendrycks et al., 2018) that considers synthetically created, labeled anomalies, we thereby use a combination of two losses that share parameters: one for the normal and one for the anomalous data. We then iteratively proceed with block coordinate updates on the parameters and the most likely (latent) labels. Our experiments with several backbone models on three image datasets, 30 tabular data sets, and a video anomaly detection benchmark showed consistent and significant improvements over the baselines.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz7FOwzAUhWEvDKjwAEz4BRJuHAe7Y5WUghSpQ9mje821sJTEletC-_aFwnSm_0ifEA8VlNo2DTxhOoWvUilQJViw9lZ0PWaes9we8xg4yfVpHw_HxNLHJFdznHA8y44zuxziLL9D_pRtnDNOYf4pP2SHGe_EjcfxwPf_uxC7l_V7-1r0281bu-oLfDa2YNAGyTEBACEujUaz1LXz3tmqUtY0WBEZYoveKfJAugZFzhI7rXW9EI9_r1fFsE9hwnQefjXDVVNfAGD-RgA</recordid><startdate>20220216</startdate><enddate>20220216</enddate><creator>Qiu, Chen</creator><creator>Li, Aodong</creator><creator>Kloft, Marius</creator><creator>Rudolph, Maja</creator><creator>Mandt, Stephan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220216</creationdate><title>Latent Outlier Exposure for Anomaly Detection with Contaminated Data</title><author>Qiu, Chen ; Li, Aodong ; Kloft, Marius ; Rudolph, Maja ; Mandt, Stephan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-e047abceb000baa974a7943cffc8112875a1bb7be8afc2bf0b4302bc8bec4443</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Qiu, Chen</creatorcontrib><creatorcontrib>Li, Aodong</creatorcontrib><creatorcontrib>Kloft, Marius</creatorcontrib><creatorcontrib>Rudolph, Maja</creatorcontrib><creatorcontrib>Mandt, Stephan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Qiu, Chen</au><au>Li, Aodong</au><au>Kloft, Marius</au><au>Rudolph, Maja</au><au>Mandt, Stephan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Latent Outlier Exposure for Anomaly Detection with Contaminated Data</atitle><date>2022-02-16</date><risdate>2022</risdate><abstract>Proceedings of the 39th International Conference on Machine Learning, 2022, volume:162, pages:18153--18167 Anomaly detection aims at identifying data points that show systematic deviations from the majority of data in an unlabeled dataset. A common assumption is that clean training data (free of anomalies) is available, which is often violated in practice. We propose a strategy for training an anomaly detector in the presence of unlabeled anomalies that is compatible with a broad class of models. The idea is to jointly infer binary labels to each datum (normal vs. anomalous) while updating the model parameters. Inspired by outlier exposure (Hendrycks et al., 2018) that considers synthetically created, labeled anomalies, we thereby use a combination of two losses that share parameters: one for the normal and one for the anomalous data. We then iteratively proceed with block coordinate updates on the parameters and the most likely (latent) labels. Our experiments with several backbone models on three image datasets, 30 tabular data sets, and a video anomaly detection benchmark showed consistent and significant improvements over the baselines.</abstract><doi>10.48550/arxiv.2202.08088</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2202.08088
ispartof
issn
language eng
recordid cdi_arxiv_primary_2202_08088
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
title Latent Outlier Exposure for Anomaly Detection with Contaminated Data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-19T14%3A22%3A25IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Latent%20Outlier%20Exposure%20for%20Anomaly%20Detection%20with%20Contaminated%20Data&rft.au=Qiu,%20Chen&rft.date=2022-02-16&rft_id=info:doi/10.48550/arxiv.2202.08088&rft_dat=%3Carxiv_GOX%3E2202_08088%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true