Transductive Learning with Prior Knowledge for Generalized Zero-shot Action Recognition

It is challenging to achieve generalized zero-shot action recognition. Different from the conventional zero-shot tasks which assume that the instances of the source classes are absent in the test set, the generalized zero-shot task studies the case that the test set contains both the source and the...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on circuits and systems for video technology 2024-01, Vol.34 (1), p.1-1
Hauptverfasser: Su, Taiyi, Wang, Hanli, Qi, Qiuping, Wang, Lei, He, Bin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1
container_issue 1
container_start_page 1
container_title IEEE transactions on circuits and systems for video technology
container_volume 34
creator Su, Taiyi
Wang, Hanli
Qi, Qiuping
Wang, Lei
He, Bin
description It is challenging to achieve generalized zero-shot action recognition. Different from the conventional zero-shot tasks which assume that the instances of the source classes are absent in the test set, the generalized zero-shot task studies the case that the test set contains both the source and the target classes. Due to the gap between visual feature and semantic embedding as well as the inherent bias of the learned classifier towards the source classes, the existing generalized zero-shot action recognition approaches are still far less effective than traditional zero-shot action recognition approaches. Facing these challenges, a novel transductive learning with prior knowledge (TLPK) model is proposed for generalized zero-shot action recognition. First, TLPK learns the prior knowledge which assists in bridging the gap between visual features and semantic embeddings, and preliminarily reduces the bias caused by the visual-semantic gap. Then, a transductive learning method that employs unlabeled target data is designed to overcome the bias problem in an effective manner. To achieve this, a target semantic-available approach and a target semantic-free approach are devised to utilize the target semantics in two different ways, where the target semantic-free approach exploits prior knowledge to produce well-performed semantic embeddings. By exploring the usage of the aforementioned prior-knowledge learning and transductive learning strategies, TLPK significantly bridges the visual-semantic gap and alleviates the bias between the source and the target classes. The experiments on the benchmark datasets of HMDB51 and UCF101 demonstrate the effectiveness of the proposed model compared to the state-of-the-art methods. The source code of this work can be found in https://mic.tongji.edu.cn.
doi_str_mv 10.1109/TCSVT.2023.3284977
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TCSVT_2023_3284977</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10147863</ieee_id><sourcerecordid>2911475727</sourcerecordid><originalsourceid>FETCH-LOGICAL-c296t-847577fb189a15b1003e79bbcabbda3c4f4b77fd79001431bd40a1cb6705c2263</originalsourceid><addsrcrecordid>eNpNUE1LAzEUDKJgrf4B8RDwvDUfm83usRStYkHRVcFLSLLZNqUmNdla9NebtT14evN4M_OGAeAcoxHGqLqqJ8-v9YggQkeUlHnF-QEYYMbKjBDEDhNGDGclwewYnMS4RAjnZc4H4K0O0sVmozv7ZeDMyOCsm8Ot7RbwMVgf4L3z25Vp5ga2aZsaZ4Jc2R_TwHcTfBYXvoPjJPcOPhnt5872-BQctXIVzdl-DsHLzXU9uc1mD9O7yXiWaVIVXZYyMM5bhctKYqYwQtTwSiktlWok1Xmbq3RveNUnplg1OZJYq4Ijpgkp6BBc7nzXwX9uTOzE0m-CSy8FqTDu7QlPLLJj6eBjDKYV62A_ZPgWGIm-QPFXoOgLFPsCk-hiJ7LGmH-CZFoWlP4CbjRtUw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2911475727</pqid></control><display><type>article</type><title>Transductive Learning with Prior Knowledge for Generalized Zero-shot Action Recognition</title><source>IEEE Electronic Library (IEL)</source><creator>Su, Taiyi ; Wang, Hanli ; Qi, Qiuping ; Wang, Lei ; He, Bin</creator><creatorcontrib>Su, Taiyi ; Wang, Hanli ; Qi, Qiuping ; Wang, Lei ; He, Bin</creatorcontrib><description>It is challenging to achieve generalized zero-shot action recognition. Different from the conventional zero-shot tasks which assume that the instances of the source classes are absent in the test set, the generalized zero-shot task studies the case that the test set contains both the source and the target classes. Due to the gap between visual feature and semantic embedding as well as the inherent bias of the learned classifier towards the source classes, the existing generalized zero-shot action recognition approaches are still far less effective than traditional zero-shot action recognition approaches. Facing these challenges, a novel transductive learning with prior knowledge (TLPK) model is proposed for generalized zero-shot action recognition. First, TLPK learns the prior knowledge which assists in bridging the gap between visual features and semantic embeddings, and preliminarily reduces the bias caused by the visual-semantic gap. Then, a transductive learning method that employs unlabeled target data is designed to overcome the bias problem in an effective manner. To achieve this, a target semantic-available approach and a target semantic-free approach are devised to utilize the target semantics in two different ways, where the target semantic-free approach exploits prior knowledge to produce well-performed semantic embeddings. By exploring the usage of the aforementioned prior-knowledge learning and transductive learning strategies, TLPK significantly bridges the visual-semantic gap and alleviates the bias between the source and the target classes. The experiments on the benchmark datasets of HMDB51 and UCF101 demonstrate the effectiveness of the proposed model compared to the state-of-the-art methods. The source code of this work can be found in https://mic.tongji.edu.cn.</description><identifier>ISSN: 1051-8215</identifier><identifier>EISSN: 1558-2205</identifier><identifier>DOI: 10.1109/TCSVT.2023.3284977</identifier><identifier>CODEN: ITCTEM</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Action recognition ; Activity recognition ; Bias ; generalized zero-shot learning ; Learning ; semantic embedding ; Semantics ; Source code ; Test sets ; transductive learning ; transfer learning</subject><ispartof>IEEE transactions on circuits and systems for video technology, 2024-01, Vol.34 (1), p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c296t-847577fb189a15b1003e79bbcabbda3c4f4b77fd79001431bd40a1cb6705c2263</citedby><cites>FETCH-LOGICAL-c296t-847577fb189a15b1003e79bbcabbda3c4f4b77fd79001431bd40a1cb6705c2263</cites><orcidid>0000-0002-4357-1095 ; 0000-0002-9999-4871 ; 0000-0003-3193-6269</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10147863$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10147863$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Su, Taiyi</creatorcontrib><creatorcontrib>Wang, Hanli</creatorcontrib><creatorcontrib>Qi, Qiuping</creatorcontrib><creatorcontrib>Wang, Lei</creatorcontrib><creatorcontrib>He, Bin</creatorcontrib><title>Transductive Learning with Prior Knowledge for Generalized Zero-shot Action Recognition</title><title>IEEE transactions on circuits and systems for video technology</title><addtitle>TCSVT</addtitle><description>It is challenging to achieve generalized zero-shot action recognition. Different from the conventional zero-shot tasks which assume that the instances of the source classes are absent in the test set, the generalized zero-shot task studies the case that the test set contains both the source and the target classes. Due to the gap between visual feature and semantic embedding as well as the inherent bias of the learned classifier towards the source classes, the existing generalized zero-shot action recognition approaches are still far less effective than traditional zero-shot action recognition approaches. Facing these challenges, a novel transductive learning with prior knowledge (TLPK) model is proposed for generalized zero-shot action recognition. First, TLPK learns the prior knowledge which assists in bridging the gap between visual features and semantic embeddings, and preliminarily reduces the bias caused by the visual-semantic gap. Then, a transductive learning method that employs unlabeled target data is designed to overcome the bias problem in an effective manner. To achieve this, a target semantic-available approach and a target semantic-free approach are devised to utilize the target semantics in two different ways, where the target semantic-free approach exploits prior knowledge to produce well-performed semantic embeddings. By exploring the usage of the aforementioned prior-knowledge learning and transductive learning strategies, TLPK significantly bridges the visual-semantic gap and alleviates the bias between the source and the target classes. The experiments on the benchmark datasets of HMDB51 and UCF101 demonstrate the effectiveness of the proposed model compared to the state-of-the-art methods. The source code of this work can be found in https://mic.tongji.edu.cn.</description><subject>Action recognition</subject><subject>Activity recognition</subject><subject>Bias</subject><subject>generalized zero-shot learning</subject><subject>Learning</subject><subject>semantic embedding</subject><subject>Semantics</subject><subject>Source code</subject><subject>Test sets</subject><subject>transductive learning</subject><subject>transfer learning</subject><issn>1051-8215</issn><issn>1558-2205</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpNUE1LAzEUDKJgrf4B8RDwvDUfm83usRStYkHRVcFLSLLZNqUmNdla9NebtT14evN4M_OGAeAcoxHGqLqqJ8-v9YggQkeUlHnF-QEYYMbKjBDEDhNGDGclwewYnMS4RAjnZc4H4K0O0sVmozv7ZeDMyOCsm8Ot7RbwMVgf4L3z25Vp5ga2aZsaZ4Jc2R_TwHcTfBYXvoPjJPcOPhnt5872-BQctXIVzdl-DsHLzXU9uc1mD9O7yXiWaVIVXZYyMM5bhctKYqYwQtTwSiktlWok1Xmbq3RveNUnplg1OZJYq4Ijpgkp6BBc7nzXwX9uTOzE0m-CSy8FqTDu7QlPLLJj6eBjDKYV62A_ZPgWGIm-QPFXoOgLFPsCk-hiJ7LGmH-CZFoWlP4CbjRtUw</recordid><startdate>20240101</startdate><enddate>20240101</enddate><creator>Su, Taiyi</creator><creator>Wang, Hanli</creator><creator>Qi, Qiuping</creator><creator>Wang, Lei</creator><creator>He, Bin</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-4357-1095</orcidid><orcidid>https://orcid.org/0000-0002-9999-4871</orcidid><orcidid>https://orcid.org/0000-0003-3193-6269</orcidid></search><sort><creationdate>20240101</creationdate><title>Transductive Learning with Prior Knowledge for Generalized Zero-shot Action Recognition</title><author>Su, Taiyi ; Wang, Hanli ; Qi, Qiuping ; Wang, Lei ; He, Bin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c296t-847577fb189a15b1003e79bbcabbda3c4f4b77fd79001431bd40a1cb6705c2263</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Action recognition</topic><topic>Activity recognition</topic><topic>Bias</topic><topic>generalized zero-shot learning</topic><topic>Learning</topic><topic>semantic embedding</topic><topic>Semantics</topic><topic>Source code</topic><topic>Test sets</topic><topic>transductive learning</topic><topic>transfer learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Su, Taiyi</creatorcontrib><creatorcontrib>Wang, Hanli</creatorcontrib><creatorcontrib>Qi, Qiuping</creatorcontrib><creatorcontrib>Wang, Lei</creatorcontrib><creatorcontrib>He, Bin</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on circuits and systems for video technology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Su, Taiyi</au><au>Wang, Hanli</au><au>Qi, Qiuping</au><au>Wang, Lei</au><au>He, Bin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Transductive Learning with Prior Knowledge for Generalized Zero-shot Action Recognition</atitle><jtitle>IEEE transactions on circuits and systems for video technology</jtitle><stitle>TCSVT</stitle><date>2024-01-01</date><risdate>2024</risdate><volume>34</volume><issue>1</issue><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>1051-8215</issn><eissn>1558-2205</eissn><coden>ITCTEM</coden><abstract>It is challenging to achieve generalized zero-shot action recognition. Different from the conventional zero-shot tasks which assume that the instances of the source classes are absent in the test set, the generalized zero-shot task studies the case that the test set contains both the source and the target classes. Due to the gap between visual feature and semantic embedding as well as the inherent bias of the learned classifier towards the source classes, the existing generalized zero-shot action recognition approaches are still far less effective than traditional zero-shot action recognition approaches. Facing these challenges, a novel transductive learning with prior knowledge (TLPK) model is proposed for generalized zero-shot action recognition. First, TLPK learns the prior knowledge which assists in bridging the gap between visual features and semantic embeddings, and preliminarily reduces the bias caused by the visual-semantic gap. Then, a transductive learning method that employs unlabeled target data is designed to overcome the bias problem in an effective manner. To achieve this, a target semantic-available approach and a target semantic-free approach are devised to utilize the target semantics in two different ways, where the target semantic-free approach exploits prior knowledge to produce well-performed semantic embeddings. By exploring the usage of the aforementioned prior-knowledge learning and transductive learning strategies, TLPK significantly bridges the visual-semantic gap and alleviates the bias between the source and the target classes. The experiments on the benchmark datasets of HMDB51 and UCF101 demonstrate the effectiveness of the proposed model compared to the state-of-the-art methods. The source code of this work can be found in https://mic.tongji.edu.cn.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TCSVT.2023.3284977</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0002-4357-1095</orcidid><orcidid>https://orcid.org/0000-0002-9999-4871</orcidid><orcidid>https://orcid.org/0000-0003-3193-6269</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1051-8215
ispartof IEEE transactions on circuits and systems for video technology, 2024-01, Vol.34 (1), p.1-1
issn 1051-8215
1558-2205
language eng
recordid cdi_crossref_primary_10_1109_TCSVT_2023_3284977
source IEEE Electronic Library (IEL)
subjects Action recognition
Activity recognition
Bias
generalized zero-shot learning
Learning
semantic embedding
Semantics
Source code
Test sets
transductive learning
transfer learning
title Transductive Learning with Prior Knowledge for Generalized Zero-shot Action Recognition
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T23%3A09%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Transductive%20Learning%20with%20Prior%20Knowledge%20for%20Generalized%20Zero-shot%20Action%20Recognition&rft.jtitle=IEEE%20transactions%20on%20circuits%20and%20systems%20for%20video%20technology&rft.au=Su,%20Taiyi&rft.date=2024-01-01&rft.volume=34&rft.issue=1&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=1051-8215&rft.eissn=1558-2205&rft.coden=ITCTEM&rft_id=info:doi/10.1109/TCSVT.2023.3284977&rft_dat=%3Cproquest_RIE%3E2911475727%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2911475727&rft_id=info:pmid/&rft_ieee_id=10147863&rfr_iscdi=true