Three-dimension Attention Mechanism And Self-supervised Pretext Task For Augmenting Few-shot Learning

The main challenge of few-shot learning lies in the limited labeled sample of data. In addition, since image-level labels are usually not accurate in describing the features of images, it leads to difficulty for the model to have good generalization ability and robustness. This problem has not been...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2023-01, Vol.11, p.1-1
Hauptverfasser: Liang, Yong, Chen, Zetao, Lin, Daoqian, Tan, Junwen, Yang, ZhenHao, Li, Jie, Li, Xinhai
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1
container_issue
container_start_page 1
container_title IEEE access
container_volume 11
creator Liang, Yong
Chen, Zetao
Lin, Daoqian
Tan, Junwen
Yang, ZhenHao
Li, Jie
Li, Xinhai
description The main challenge of few-shot learning lies in the limited labeled sample of data. In addition, since image-level labels are usually not accurate in describing the features of images, it leads to difficulty for the model to have good generalization ability and robustness. This problem has not been well solved yet, and existing metric-based methods still have room for improvement. To address this issue, we propose a few-shot learning method based on a three-dimension attention mechanism and self-supervised learning. The attention module is used to extract more representative features by focusing on more semantically informative features through spatial and channel attention. Self-supervised learning mainly adopts a proxy task of rotation transformation, which increases semantic information without requiring additional manual labeling, and uses this information for training in combination with supervised learning loss function to improve model robustness. We have conducted extensive experiments on four popular few-shot datasets and achieved state-of-the-art performance in both 5-shot and 1-shot scenarios. Experiment results show that our work provides a novel and remarkable approach to few-shot learning.
doi_str_mv 10.1109/ACCESS.2023.3285721
format Article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_10149320</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10149320</ieee_id><doaj_id>oai_doaj_org_article_68d72b58d95442bd80186029ad5da3fc</doaj_id><sourcerecordid>2828008610</sourcerecordid><originalsourceid>FETCH-LOGICAL-c359t-7edcb7a16ea964c71da59e1f7d45068abc9094cd6322ae60f5dc2805fc9739693</originalsourceid><addsrcrecordid>eNpNUctKAzEUHURB0X6BLgKup-YxySTLobRaqCi0rkOa3GmntpOapD7-3qkj4t3cy-E8LpwsuyZ4SAhWd9VoNJ7PhxRTNmRU8pKSk-yCEqFyxpk4_XefZ4MYN7gb2UG8vMhgsQ4AuWt20MbGt6hKCdp0vB7Brk3bxB2qWofmsK3zeNhDeG8iOPQcIMFnQgsTX9HEB1QdVrujsl2hCXzkce0TmoEJbYdcZWe12UYY_O7L7GUyXowe8tnT_XRUzXLLuEp5Cc4uS0MEGCUKWxJnuAJSl67gWEiztAqrwjrBKDUgcM2dpRLz2qqSKaHYZTbtfZ03G70Pzc6EL-1No38AH1bahNTYLWghXUmXXDrFi4IuncRECkyVcdwZVtvO67b32gf_doCY9MYfQtu9r6nsUrEUBHcs1rNs8DEGqP9SCdbHenRfjz7Wo3_r6VQ3vaoBgH8KUihGMfsGIgSLyA</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2828008610</pqid></control><display><type>article</type><title>Three-dimension Attention Mechanism And Self-supervised Pretext Task For Augmenting Few-shot Learning</title><source>DOAJ Directory of Open Access Journals</source><source>IEEE Xplore Open Access Journals</source><source>EZB Electronic Journals Library</source><creator>Liang, Yong ; Chen, Zetao ; Lin, Daoqian ; Tan, Junwen ; Yang, ZhenHao ; Li, Jie ; Li, Xinhai</creator><creatorcontrib>Liang, Yong ; Chen, Zetao ; Lin, Daoqian ; Tan, Junwen ; Yang, ZhenHao ; Li, Jie ; Li, Xinhai</creatorcontrib><description>The main challenge of few-shot learning lies in the limited labeled sample of data. In addition, since image-level labels are usually not accurate in describing the features of images, it leads to difficulty for the model to have good generalization ability and robustness. This problem has not been well solved yet, and existing metric-based methods still have room for improvement. To address this issue, we propose a few-shot learning method based on a three-dimension attention mechanism and self-supervised learning. The attention module is used to extract more representative features by focusing on more semantically informative features through spatial and channel attention. Self-supervised learning mainly adopts a proxy task of rotation transformation, which increases semantic information without requiring additional manual labeling, and uses this information for training in combination with supervised learning loss function to improve model robustness. We have conducted extensive experiments on four popular few-shot datasets and achieved state-of-the-art performance in both 5-shot and 1-shot scenarios. Experiment results show that our work provides a novel and remarkable approach to few-shot learning.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2023.3285721</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Attention mechanism ; Data mining ; Deep learning ; Feature extraction ; Few-shot ; Image classification ; Labels ; Robustness ; Self-supervised learning ; Self-supervised pretext task learning ; Supervised learning</subject><ispartof>IEEE access, 2023-01, Vol.11, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c359t-7edcb7a16ea964c71da59e1f7d45068abc9094cd6322ae60f5dc2805fc9739693</cites><orcidid>0000-0002-5658-2260</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10149320$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,860,2095,27612,27903,27904,54912</link.rule.ids></links><search><creatorcontrib>Liang, Yong</creatorcontrib><creatorcontrib>Chen, Zetao</creatorcontrib><creatorcontrib>Lin, Daoqian</creatorcontrib><creatorcontrib>Tan, Junwen</creatorcontrib><creatorcontrib>Yang, ZhenHao</creatorcontrib><creatorcontrib>Li, Jie</creatorcontrib><creatorcontrib>Li, Xinhai</creatorcontrib><title>Three-dimension Attention Mechanism And Self-supervised Pretext Task For Augmenting Few-shot Learning</title><title>IEEE access</title><addtitle>Access</addtitle><description>The main challenge of few-shot learning lies in the limited labeled sample of data. In addition, since image-level labels are usually not accurate in describing the features of images, it leads to difficulty for the model to have good generalization ability and robustness. This problem has not been well solved yet, and existing metric-based methods still have room for improvement. To address this issue, we propose a few-shot learning method based on a three-dimension attention mechanism and self-supervised learning. The attention module is used to extract more representative features by focusing on more semantically informative features through spatial and channel attention. Self-supervised learning mainly adopts a proxy task of rotation transformation, which increases semantic information without requiring additional manual labeling, and uses this information for training in combination with supervised learning loss function to improve model robustness. We have conducted extensive experiments on four popular few-shot datasets and achieved state-of-the-art performance in both 5-shot and 1-shot scenarios. Experiment results show that our work provides a novel and remarkable approach to few-shot learning.</description><subject>Attention mechanism</subject><subject>Data mining</subject><subject>Deep learning</subject><subject>Feature extraction</subject><subject>Few-shot</subject><subject>Image classification</subject><subject>Labels</subject><subject>Robustness</subject><subject>Self-supervised learning</subject><subject>Self-supervised pretext task learning</subject><subject>Supervised learning</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNUctKAzEUHURB0X6BLgKup-YxySTLobRaqCi0rkOa3GmntpOapD7-3qkj4t3cy-E8LpwsuyZ4SAhWd9VoNJ7PhxRTNmRU8pKSk-yCEqFyxpk4_XefZ4MYN7gb2UG8vMhgsQ4AuWt20MbGt6hKCdp0vB7Brk3bxB2qWofmsK3zeNhDeG8iOPQcIMFnQgsTX9HEB1QdVrujsl2hCXzkce0TmoEJbYdcZWe12UYY_O7L7GUyXowe8tnT_XRUzXLLuEp5Cc4uS0MEGCUKWxJnuAJSl67gWEiztAqrwjrBKDUgcM2dpRLz2qqSKaHYZTbtfZ03G70Pzc6EL-1No38AH1bahNTYLWghXUmXXDrFi4IuncRECkyVcdwZVtvO67b32gf_doCY9MYfQtu9r6nsUrEUBHcs1rNs8DEGqP9SCdbHenRfjz7Wo3_r6VQ3vaoBgH8KUihGMfsGIgSLyA</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Liang, Yong</creator><creator>Chen, Zetao</creator><creator>Lin, Daoqian</creator><creator>Tan, Junwen</creator><creator>Yang, ZhenHao</creator><creator>Li, Jie</creator><creator>Li, Xinhai</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-5658-2260</orcidid></search><sort><creationdate>20230101</creationdate><title>Three-dimension Attention Mechanism And Self-supervised Pretext Task For Augmenting Few-shot Learning</title><author>Liang, Yong ; Chen, Zetao ; Lin, Daoqian ; Tan, Junwen ; Yang, ZhenHao ; Li, Jie ; Li, Xinhai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c359t-7edcb7a16ea964c71da59e1f7d45068abc9094cd6322ae60f5dc2805fc9739693</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Attention mechanism</topic><topic>Data mining</topic><topic>Deep learning</topic><topic>Feature extraction</topic><topic>Few-shot</topic><topic>Image classification</topic><topic>Labels</topic><topic>Robustness</topic><topic>Self-supervised learning</topic><topic>Self-supervised pretext task learning</topic><topic>Supervised learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Liang, Yong</creatorcontrib><creatorcontrib>Chen, Zetao</creatorcontrib><creatorcontrib>Lin, Daoqian</creatorcontrib><creatorcontrib>Tan, Junwen</creatorcontrib><creatorcontrib>Yang, ZhenHao</creatorcontrib><creatorcontrib>Li, Jie</creatorcontrib><creatorcontrib>Li, Xinhai</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998–Present</collection><collection>IEEE Xplore (Online service)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Liang, Yong</au><au>Chen, Zetao</au><au>Lin, Daoqian</au><au>Tan, Junwen</au><au>Yang, ZhenHao</au><au>Li, Jie</au><au>Li, Xinhai</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Three-dimension Attention Mechanism And Self-supervised Pretext Task For Augmenting Few-shot Learning</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2023-01-01</date><risdate>2023</risdate><volume>11</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>The main challenge of few-shot learning lies in the limited labeled sample of data. In addition, since image-level labels are usually not accurate in describing the features of images, it leads to difficulty for the model to have good generalization ability and robustness. This problem has not been well solved yet, and existing metric-based methods still have room for improvement. To address this issue, we propose a few-shot learning method based on a three-dimension attention mechanism and self-supervised learning. The attention module is used to extract more representative features by focusing on more semantically informative features through spatial and channel attention. Self-supervised learning mainly adopts a proxy task of rotation transformation, which increases semantic information without requiring additional manual labeling, and uses this information for training in combination with supervised learning loss function to improve model robustness. We have conducted extensive experiments on four popular few-shot datasets and achieved state-of-the-art performance in both 5-shot and 1-shot scenarios. Experiment results show that our work provides a novel and remarkable approach to few-shot learning.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2023.3285721</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0002-5658-2260</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2023-01, Vol.11, p.1-1
issn 2169-3536
2169-3536
language eng
recordid cdi_ieee_primary_10149320
source DOAJ Directory of Open Access Journals; IEEE Xplore Open Access Journals; EZB Electronic Journals Library
subjects Attention mechanism
Data mining
Deep learning
Feature extraction
Few-shot
Image classification
Labels
Robustness
Self-supervised learning
Self-supervised pretext task learning
Supervised learning
title Three-dimension Attention Mechanism And Self-supervised Pretext Task For Augmenting Few-shot Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-22T18%3A27%3A38IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Three-dimension%20Attention%20Mechanism%20And%20Self-supervised%20Pretext%20Task%20For%20Augmenting%20Few-shot%20Learning&rft.jtitle=IEEE%20access&rft.au=Liang,%20Yong&rft.date=2023-01-01&rft.volume=11&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2023.3285721&rft_dat=%3Cproquest_ieee_%3E2828008610%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2828008610&rft_id=info:pmid/&rft_ieee_id=10149320&rft_doaj_id=oai_doaj_org_article_68d72b58d95442bd80186029ad5da3fc&rfr_iscdi=true