OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features
In this paper, we consider the task of one-shot object detection, which consists in detecting objects defined by a single demonstration. Differently from the standard object detection, the classes of objects used for training and testing do not overlap. We build the one-stage system that performs lo...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2020-08 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Osokin, Anton Sumin, Denis Lomakin, Vasily |
description | In this paper, we consider the task of one-shot object detection, which consists in detecting objects defined by a single demonstration. Differently from the standard object detection, the classes of objects used for training and testing do not overlap. We build the one-stage system that performs localization and recognition jointly. We use dense correlation matching of learned local features to find correspondences, a feed-forward geometric transformation model to align features and bilinear resampling of the correlation tensor to compute the detection score of the aligned features. All the components are differentiable, which allows end-to-end training. Experimental evaluation on several challenging domains (retail products, 3D objects, buildings and logos) shows that our method can detect unseen classes (e.g., toothpaste when trained on groceries) and outperforms several baselines by a significant margin. Our code is available online: https://github.com/aosokin/os2d . |
format | Article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2377808676</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2377808676</sourcerecordid><originalsourceid>FETCH-proquest_journals_23778086763</originalsourceid><addsrcrecordid>eNqNissKgkAUQIcgSMp_uNBamGbUkXaRSptwYXsZ5eaDmKmZ66K_T6gPaHUOnLNigZDyEGWxEBsWej9xzkWqRJLIgBVVLfIjVAajmnSPXxssQdVO2BHkSAtGa6B9w1VTN4ymh5PpBuugRE2zQ79j67t-eAx_3LJ9WdzOl-jp7GtGT81kZ2eW1AipVMazVKXyv-sDgrA5Tg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2377808676</pqid></control><display><type>article</type><title>OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features</title><source>Free E- Journals</source><creator>Osokin, Anton ; Sumin, Denis ; Lomakin, Vasily</creator><creatorcontrib>Osokin, Anton ; Sumin, Denis ; Lomakin, Vasily</creatorcontrib><description>In this paper, we consider the task of one-shot object detection, which consists in detecting objects defined by a single demonstration. Differently from the standard object detection, the classes of objects used for training and testing do not overlap. We build the one-stage system that performs localization and recognition jointly. We use dense correlation matching of learned local features to find correspondences, a feed-forward geometric transformation model to align features and bilinear resampling of the correlation tensor to compute the detection score of the aligned features. All the components are differentiable, which allows end-to-end training. Experimental evaluation on several challenging domains (retail products, 3D objects, buildings and logos) shows that our method can detect unseen classes (e.g., toothpaste when trained on groceries) and outperforms several baselines by a significant margin. Our code is available online: https://github.com/aosokin/os2d .</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Geometric transformation ; Groceries ; Logos ; Matching ; Object recognition ; Resampling ; Tensors ; Toothpaste ; Training</subject><ispartof>arXiv.org, 2020-08</ispartof><rights>2020. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>777,781</link.rule.ids></links><search><creatorcontrib>Osokin, Anton</creatorcontrib><creatorcontrib>Sumin, Denis</creatorcontrib><creatorcontrib>Lomakin, Vasily</creatorcontrib><title>OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features</title><title>arXiv.org</title><description>In this paper, we consider the task of one-shot object detection, which consists in detecting objects defined by a single demonstration. Differently from the standard object detection, the classes of objects used for training and testing do not overlap. We build the one-stage system that performs localization and recognition jointly. We use dense correlation matching of learned local features to find correspondences, a feed-forward geometric transformation model to align features and bilinear resampling of the correlation tensor to compute the detection score of the aligned features. All the components are differentiable, which allows end-to-end training. Experimental evaluation on several challenging domains (retail products, 3D objects, buildings and logos) shows that our method can detect unseen classes (e.g., toothpaste when trained on groceries) and outperforms several baselines by a significant margin. Our code is available online: https://github.com/aosokin/os2d .</description><subject>Geometric transformation</subject><subject>Groceries</subject><subject>Logos</subject><subject>Matching</subject><subject>Object recognition</subject><subject>Resampling</subject><subject>Tensors</subject><subject>Toothpaste</subject><subject>Training</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNissKgkAUQIcgSMp_uNBamGbUkXaRSptwYXsZ5eaDmKmZ66K_T6gPaHUOnLNigZDyEGWxEBsWej9xzkWqRJLIgBVVLfIjVAajmnSPXxssQdVO2BHkSAtGa6B9w1VTN4ymh5PpBuugRE2zQ79j67t-eAx_3LJ9WdzOl-jp7GtGT81kZ2eW1AipVMazVKXyv-sDgrA5Tg</recordid><startdate>20200819</startdate><enddate>20200819</enddate><creator>Osokin, Anton</creator><creator>Sumin, Denis</creator><creator>Lomakin, Vasily</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20200819</creationdate><title>OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features</title><author>Osokin, Anton ; Sumin, Denis ; Lomakin, Vasily</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_23778086763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Geometric transformation</topic><topic>Groceries</topic><topic>Logos</topic><topic>Matching</topic><topic>Object recognition</topic><topic>Resampling</topic><topic>Tensors</topic><topic>Toothpaste</topic><topic>Training</topic><toplevel>online_resources</toplevel><creatorcontrib>Osokin, Anton</creatorcontrib><creatorcontrib>Sumin, Denis</creatorcontrib><creatorcontrib>Lomakin, Vasily</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Osokin, Anton</au><au>Sumin, Denis</au><au>Lomakin, Vasily</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features</atitle><jtitle>arXiv.org</jtitle><date>2020-08-19</date><risdate>2020</risdate><eissn>2331-8422</eissn><abstract>In this paper, we consider the task of one-shot object detection, which consists in detecting objects defined by a single demonstration. Differently from the standard object detection, the classes of objects used for training and testing do not overlap. We build the one-stage system that performs localization and recognition jointly. We use dense correlation matching of learned local features to find correspondences, a feed-forward geometric transformation model to align features and bilinear resampling of the correlation tensor to compute the detection score of the aligned features. All the components are differentiable, which allows end-to-end training. Experimental evaluation on several challenging domains (retail products, 3D objects, buildings and logos) shows that our method can detect unseen classes (e.g., toothpaste when trained on groceries) and outperforms several baselines by a significant margin. Our code is available online: https://github.com/aosokin/os2d .</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2020-08 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2377808676 |
source | Free E- Journals |
subjects | Geometric transformation Groceries Logos Matching Object recognition Resampling Tensors Toothpaste Training |
title | OS2D: One-Stage One-Shot Object Detection by Matching Anchor Features |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-18T11%3A21%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=OS2D:%20One-Stage%20One-Shot%20Object%20Detection%20by%20Matching%20Anchor%20Features&rft.jtitle=arXiv.org&rft.au=Osokin,%20Anton&rft.date=2020-08-19&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2377808676%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2377808676&rft_id=info:pmid/&rfr_iscdi=true |