LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning

We present a novel vision-language prompt learning approach for few-shot out-of-distribution (OOD) detection. Few-shot OOD detection aims to detect OOD images from classes that are unseen during training using only a few labeled in-distribution (ID) images. While prompt learning methods such as CoOp...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Miyai, Atsuyuki, Yu, Qing, Irie, Go, Aizawa, Kiyoharu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Miyai, Atsuyuki
Yu, Qing
Irie, Go
Aizawa, Kiyoharu
description We present a novel vision-language prompt learning approach for few-shot out-of-distribution (OOD) detection. Few-shot OOD detection aims to detect OOD images from classes that are unseen during training using only a few labeled in-distribution (ID) images. While prompt learning methods such as CoOp have shown effectiveness and efficiency in few-shot ID classification, they still face limitations in OOD detection due to the potential presence of ID-irrelevant information in text embeddings. To address this issue, we introduce a new approach called Local regularized Context Optimization (LoCoOp), which performs OOD regularization that utilizes the portions of CLIP local features as OOD features during training. CLIP's local features have a lot of ID-irrelevant nuisances (e.g., backgrounds), and by learning to push them away from the ID class text embeddings, we can remove the nuisances in the ID class text embeddings and enhance the separation between ID and OOD. Experiments on the large-scale ImageNet OOD detection benchmarks demonstrate the superiority of our LoCoOp over zero-shot, fully supervised detection methods and prompt learning methods. Notably, even in a one-shot setting -- just one label per class, LoCoOp outperforms existing zero-shot and fully supervised detection methods. The code will be available via https://github.com/AtsuMiyai/LoCoOp.
doi_str_mv 10.48550/arxiv.2306.01293
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2306_01293</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2306_01293</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-5944b38e1e444062625f86b665b3e482597f48f599f26198ea99d3959f094f393</originalsourceid><addsrcrecordid>eNotz7FOwzAUhWEvDKjwAEz4BRwc32vXlw1SCkiRgtTukVNssETjyHULvD0iMJ1_OtLH2FUtK7RayxuXv-KpUiBNJWtFcM7u29Skbrrla_8pNu-p8O5YRApiFQ8lx-FYYhr5yhe_m-sUHX_JaT8V3nqXxzi-XbCz4D4O_vJ_F2y7ftg2T6LtHp-bu1Y4swShCXEA62uPiNIoo3SwZjBGD-DRKk3LgDZooqBMTdY7olcgTUESBiBYsOu_2xnRTznuXf7ufzH9jIEfyGNCYg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning</title><source>arXiv.org</source><creator>Miyai, Atsuyuki ; Yu, Qing ; Irie, Go ; Aizawa, Kiyoharu</creator><creatorcontrib>Miyai, Atsuyuki ; Yu, Qing ; Irie, Go ; Aizawa, Kiyoharu</creatorcontrib><description>We present a novel vision-language prompt learning approach for few-shot out-of-distribution (OOD) detection. Few-shot OOD detection aims to detect OOD images from classes that are unseen during training using only a few labeled in-distribution (ID) images. While prompt learning methods such as CoOp have shown effectiveness and efficiency in few-shot ID classification, they still face limitations in OOD detection due to the potential presence of ID-irrelevant information in text embeddings. To address this issue, we introduce a new approach called Local regularized Context Optimization (LoCoOp), which performs OOD regularization that utilizes the portions of CLIP local features as OOD features during training. CLIP's local features have a lot of ID-irrelevant nuisances (e.g., backgrounds), and by learning to push them away from the ID class text embeddings, we can remove the nuisances in the ID class text embeddings and enhance the separation between ID and OOD. Experiments on the large-scale ImageNet OOD detection benchmarks demonstrate the superiority of our LoCoOp over zero-shot, fully supervised detection methods and prompt learning methods. Notably, even in a one-shot setting -- just one label per class, LoCoOp outperforms existing zero-shot and fully supervised detection methods. The code will be available via https://github.com/AtsuMiyai/LoCoOp.</description><identifier>DOI: 10.48550/arxiv.2306.01293</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2023-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2306.01293$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2306.01293$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Miyai, Atsuyuki</creatorcontrib><creatorcontrib>Yu, Qing</creatorcontrib><creatorcontrib>Irie, Go</creatorcontrib><creatorcontrib>Aizawa, Kiyoharu</creatorcontrib><title>LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning</title><description>We present a novel vision-language prompt learning approach for few-shot out-of-distribution (OOD) detection. Few-shot OOD detection aims to detect OOD images from classes that are unseen during training using only a few labeled in-distribution (ID) images. While prompt learning methods such as CoOp have shown effectiveness and efficiency in few-shot ID classification, they still face limitations in OOD detection due to the potential presence of ID-irrelevant information in text embeddings. To address this issue, we introduce a new approach called Local regularized Context Optimization (LoCoOp), which performs OOD regularization that utilizes the portions of CLIP local features as OOD features during training. CLIP's local features have a lot of ID-irrelevant nuisances (e.g., backgrounds), and by learning to push them away from the ID class text embeddings, we can remove the nuisances in the ID class text embeddings and enhance the separation between ID and OOD. Experiments on the large-scale ImageNet OOD detection benchmarks demonstrate the superiority of our LoCoOp over zero-shot, fully supervised detection methods and prompt learning methods. Notably, even in a one-shot setting -- just one label per class, LoCoOp outperforms existing zero-shot and fully supervised detection methods. The code will be available via https://github.com/AtsuMiyai/LoCoOp.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz7FOwzAUhWEvDKjwAEz4BRwc32vXlw1SCkiRgtTukVNssETjyHULvD0iMJ1_OtLH2FUtK7RayxuXv-KpUiBNJWtFcM7u29Skbrrla_8pNu-p8O5YRApiFQ8lx-FYYhr5yhe_m-sUHX_JaT8V3nqXxzi-XbCz4D4O_vJ_F2y7ftg2T6LtHp-bu1Y4swShCXEA62uPiNIoo3SwZjBGD-DRKk3LgDZooqBMTdY7olcgTUESBiBYsOu_2xnRTznuXf7ufzH9jIEfyGNCYg</recordid><startdate>20230602</startdate><enddate>20230602</enddate><creator>Miyai, Atsuyuki</creator><creator>Yu, Qing</creator><creator>Irie, Go</creator><creator>Aizawa, Kiyoharu</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230602</creationdate><title>LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning</title><author>Miyai, Atsuyuki ; Yu, Qing ; Irie, Go ; Aizawa, Kiyoharu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-5944b38e1e444062625f86b665b3e482597f48f599f26198ea99d3959f094f393</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Miyai, Atsuyuki</creatorcontrib><creatorcontrib>Yu, Qing</creatorcontrib><creatorcontrib>Irie, Go</creatorcontrib><creatorcontrib>Aizawa, Kiyoharu</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Miyai, Atsuyuki</au><au>Yu, Qing</au><au>Irie, Go</au><au>Aizawa, Kiyoharu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning</atitle><date>2023-06-02</date><risdate>2023</risdate><abstract>We present a novel vision-language prompt learning approach for few-shot out-of-distribution (OOD) detection. Few-shot OOD detection aims to detect OOD images from classes that are unseen during training using only a few labeled in-distribution (ID) images. While prompt learning methods such as CoOp have shown effectiveness and efficiency in few-shot ID classification, they still face limitations in OOD detection due to the potential presence of ID-irrelevant information in text embeddings. To address this issue, we introduce a new approach called Local regularized Context Optimization (LoCoOp), which performs OOD regularization that utilizes the portions of CLIP local features as OOD features during training. CLIP's local features have a lot of ID-irrelevant nuisances (e.g., backgrounds), and by learning to push them away from the ID class text embeddings, we can remove the nuisances in the ID class text embeddings and enhance the separation between ID and OOD. Experiments on the large-scale ImageNet OOD detection benchmarks demonstrate the superiority of our LoCoOp over zero-shot, fully supervised detection methods and prompt learning methods. Notably, even in a one-shot setting -- just one label per class, LoCoOp outperforms existing zero-shot and fully supervised detection methods. The code will be available via https://github.com/AtsuMiyai/LoCoOp.</abstract><doi>10.48550/arxiv.2306.01293</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2306.01293
ispartof
issn
language eng
recordid cdi_arxiv_primary_2306_01293
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T08%3A13%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=LoCoOp:%20Few-Shot%20Out-of-Distribution%20Detection%20via%20Prompt%20Learning&rft.au=Miyai,%20Atsuyuki&rft.date=2023-06-02&rft_id=info:doi/10.48550/arxiv.2306.01293&rft_dat=%3Carxiv_GOX%3E2306_01293%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true