Generalized Zero-Shot Image Classification via Partially-Shared Multi-Task Representation Learning
Generalized Zero-Shot Learning (GZSL) holds significant research importance as it enables the classification of samples from both seen and unseen classes. A prevailing approach for GZSL is learning transferable representations that can generalize well to both seen and unseen classes during testing....
Gespeichert in:
Veröffentlicht in: | Electronics (Basel) 2023-05, Vol.12 (9), p.2085 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 9 |
container_start_page | 2085 |
container_title | Electronics (Basel) |
container_volume | 12 |
creator | Wang, Gerui Tang, Sheng |
description | Generalized Zero-Shot Learning (GZSL) holds significant research importance as it enables the classification of samples from both seen and unseen classes. A prevailing approach for GZSL is learning transferable representations that can generalize well to both seen and unseen classes during testing. This approach encompasses two key concepts: discriminative representations and semantic-relevant representations. “Semantic-relevant” facilitates the transfer of semantic knowledge using pre-defined semantic descriptors, while “discriminative” is crucial for accurate category discrimination. However, these two concepts are arguably inherently conflicting, as semantic descriptors are not specifically designed for image classification. Existing methods often struggle with balancing these two aspects and neglect the conflict between them, leading to suboptimal representation generalization and transferability to unseen classes. To address this issue, we propose a novel partially-shared multi-task representation learning method, termed PS-GZSL, which jointly preserves complementary and sharable knowledge between these two concepts. Specifically, we first propose a novel perspective that treats the learning of discriminative and semantic-relevant representations as optimizing a discrimination task and a visual-semantic alignment task, respectively. Then, to learn more complete and generalizable representations, PS-GZSL explicitly factorizes visual features into task-shared and task-specific representations and introduces two advanced tasks: an instance-level contrastive discrimination task and a relation-based visual-semantic alignment task. Furthermore, PS-GZSL employs Mixture-of-Experts (MoE) with a dropout mechanism to prevent representation degeneration and integrates a conditional GAN (cGAN) to synthesize unseen features for estimating unseen visual features. Extensive experiments and more competitive results on five widely-used GZSL benchmark datasets validate the effectiveness of our PS-GZSL. |
doi_str_mv | 10.3390/electronics12092085 |
format | Article |
fullrecord | <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_journals_2812386848</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A749097176</galeid><sourcerecordid>A749097176</sourcerecordid><originalsourceid>FETCH-LOGICAL-c311t-adcee00560ddfdd86be7f5c51de22b96b2ee7363677ea4bcc98493a0dde6a5473</originalsourceid><addsrcrecordid>eNptkT9PwzAQxSMEElXpJ2CJxJziP0kcj1UFpVIRCMrCEl2cS3FJ7WK7SOXTYxQGBu6GO51-793wkuSSkinnklxjjyo4a7TylBHJSFWcJCNGhMwkk-z0z36eTLzfkliS8oqTUdIs0KCDXn9hm76is9nzmw3pcgcbTOc9eK87rSBoa9JPDekjuKCh74-RAxc194c-6GwN_j19wr1DjyYM-ArBGW02F8lZB73Hye8cJy-3N-v5XbZ6WCzns1WmOKUhg1YhElKUpG27tq3KBkVXqIK2yFgjy4YhCl7yUgiEvFFKVrnkEGksocgFHydXg-_e2Y8D-lBv7cGZ-LJmFWW8Kqu8itR0oDbQY61NZ4MDFbvFnVbWYKfjfSZySaSgoowCPgiUs9477Oq90ztwx5qS-ieA-p8A-Dcud33N</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2812386848</pqid></control><display><type>article</type><title>Generalized Zero-Shot Image Classification via Partially-Shared Multi-Task Representation Learning</title><source>MDPI - Multidisciplinary Digital Publishing Institute</source><source>EZB-FREE-00999 freely available EZB journals</source><creator>Wang, Gerui ; Tang, Sheng</creator><creatorcontrib>Wang, Gerui ; Tang, Sheng</creatorcontrib><description>Generalized Zero-Shot Learning (GZSL) holds significant research importance as it enables the classification of samples from both seen and unseen classes. A prevailing approach for GZSL is learning transferable representations that can generalize well to both seen and unseen classes during testing. This approach encompasses two key concepts: discriminative representations and semantic-relevant representations. “Semantic-relevant” facilitates the transfer of semantic knowledge using pre-defined semantic descriptors, while “discriminative” is crucial for accurate category discrimination. However, these two concepts are arguably inherently conflicting, as semantic descriptors are not specifically designed for image classification. Existing methods often struggle with balancing these two aspects and neglect the conflict between them, leading to suboptimal representation generalization and transferability to unseen classes. To address this issue, we propose a novel partially-shared multi-task representation learning method, termed PS-GZSL, which jointly preserves complementary and sharable knowledge between these two concepts. Specifically, we first propose a novel perspective that treats the learning of discriminative and semantic-relevant representations as optimizing a discrimination task and a visual-semantic alignment task, respectively. Then, to learn more complete and generalizable representations, PS-GZSL explicitly factorizes visual features into task-shared and task-specific representations and introduces two advanced tasks: an instance-level contrastive discrimination task and a relation-based visual-semantic alignment task. Furthermore, PS-GZSL employs Mixture-of-Experts (MoE) with a dropout mechanism to prevent representation degeneration and integrates a conditional GAN (cGAN) to synthesize unseen features for estimating unseen visual features. Extensive experiments and more competitive results on five widely-used GZSL benchmark datasets validate the effectiveness of our PS-GZSL.</description><identifier>ISSN: 2079-9292</identifier><identifier>EISSN: 2079-9292</identifier><identifier>DOI: 10.3390/electronics12092085</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Alignment ; Automatic classification ; Classification ; Degeneration ; Image classification ; Image processing ; Knowledge management ; Learning ; Machine learning ; Methods ; Neural networks ; Representations ; Semantics ; Visual discrimination ; Visual tasks ; Zero-shot learning</subject><ispartof>Electronics (Basel), 2023-05, Vol.12 (9), p.2085</ispartof><rights>COPYRIGHT 2023 MDPI AG</rights><rights>2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c311t-adcee00560ddfdd86be7f5c51de22b96b2ee7363677ea4bcc98493a0dde6a5473</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,778,782,27911,27912</link.rule.ids></links><search><creatorcontrib>Wang, Gerui</creatorcontrib><creatorcontrib>Tang, Sheng</creatorcontrib><title>Generalized Zero-Shot Image Classification via Partially-Shared Multi-Task Representation Learning</title><title>Electronics (Basel)</title><description>Generalized Zero-Shot Learning (GZSL) holds significant research importance as it enables the classification of samples from both seen and unseen classes. A prevailing approach for GZSL is learning transferable representations that can generalize well to both seen and unseen classes during testing. This approach encompasses two key concepts: discriminative representations and semantic-relevant representations. “Semantic-relevant” facilitates the transfer of semantic knowledge using pre-defined semantic descriptors, while “discriminative” is crucial for accurate category discrimination. However, these two concepts are arguably inherently conflicting, as semantic descriptors are not specifically designed for image classification. Existing methods often struggle with balancing these two aspects and neglect the conflict between them, leading to suboptimal representation generalization and transferability to unseen classes. To address this issue, we propose a novel partially-shared multi-task representation learning method, termed PS-GZSL, which jointly preserves complementary and sharable knowledge between these two concepts. Specifically, we first propose a novel perspective that treats the learning of discriminative and semantic-relevant representations as optimizing a discrimination task and a visual-semantic alignment task, respectively. Then, to learn more complete and generalizable representations, PS-GZSL explicitly factorizes visual features into task-shared and task-specific representations and introduces two advanced tasks: an instance-level contrastive discrimination task and a relation-based visual-semantic alignment task. Furthermore, PS-GZSL employs Mixture-of-Experts (MoE) with a dropout mechanism to prevent representation degeneration and integrates a conditional GAN (cGAN) to synthesize unseen features for estimating unseen visual features. Extensive experiments and more competitive results on five widely-used GZSL benchmark datasets validate the effectiveness of our PS-GZSL.</description><subject>Alignment</subject><subject>Automatic classification</subject><subject>Classification</subject><subject>Degeneration</subject><subject>Image classification</subject><subject>Image processing</subject><subject>Knowledge management</subject><subject>Learning</subject><subject>Machine learning</subject><subject>Methods</subject><subject>Neural networks</subject><subject>Representations</subject><subject>Semantics</subject><subject>Visual discrimination</subject><subject>Visual tasks</subject><subject>Zero-shot learning</subject><issn>2079-9292</issn><issn>2079-9292</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNptkT9PwzAQxSMEElXpJ2CJxJziP0kcj1UFpVIRCMrCEl2cS3FJ7WK7SOXTYxQGBu6GO51-793wkuSSkinnklxjjyo4a7TylBHJSFWcJCNGhMwkk-z0z36eTLzfkliS8oqTUdIs0KCDXn9hm76is9nzmw3pcgcbTOc9eK87rSBoa9JPDekjuKCh74-RAxc194c-6GwN_j19wr1DjyYM-ArBGW02F8lZB73Hye8cJy-3N-v5XbZ6WCzns1WmOKUhg1YhElKUpG27tq3KBkVXqIK2yFgjy4YhCl7yUgiEvFFKVrnkEGksocgFHydXg-_e2Y8D-lBv7cGZ-LJmFWW8Kqu8itR0oDbQY61NZ4MDFbvFnVbWYKfjfSZySaSgoowCPgiUs9477Oq90ztwx5qS-ieA-p8A-Dcud33N</recordid><startdate>20230503</startdate><enddate>20230503</enddate><creator>Wang, Gerui</creator><creator>Tang, Sheng</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SP</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L7M</scope><scope>P5Z</scope><scope>P62</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope></search><sort><creationdate>20230503</creationdate><title>Generalized Zero-Shot Image Classification via Partially-Shared Multi-Task Representation Learning</title><author>Wang, Gerui ; Tang, Sheng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c311t-adcee00560ddfdd86be7f5c51de22b96b2ee7363677ea4bcc98493a0dde6a5473</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Alignment</topic><topic>Automatic classification</topic><topic>Classification</topic><topic>Degeneration</topic><topic>Image classification</topic><topic>Image processing</topic><topic>Knowledge management</topic><topic>Learning</topic><topic>Machine learning</topic><topic>Methods</topic><topic>Neural networks</topic><topic>Representations</topic><topic>Semantics</topic><topic>Visual discrimination</topic><topic>Visual tasks</topic><topic>Zero-shot learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Gerui</creatorcontrib><creatorcontrib>Tang, Sheng</creatorcontrib><collection>CrossRef</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><jtitle>Electronics (Basel)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Gerui</au><au>Tang, Sheng</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Generalized Zero-Shot Image Classification via Partially-Shared Multi-Task Representation Learning</atitle><jtitle>Electronics (Basel)</jtitle><date>2023-05-03</date><risdate>2023</risdate><volume>12</volume><issue>9</issue><spage>2085</spage><pages>2085-</pages><issn>2079-9292</issn><eissn>2079-9292</eissn><abstract>Generalized Zero-Shot Learning (GZSL) holds significant research importance as it enables the classification of samples from both seen and unseen classes. A prevailing approach for GZSL is learning transferable representations that can generalize well to both seen and unseen classes during testing. This approach encompasses two key concepts: discriminative representations and semantic-relevant representations. “Semantic-relevant” facilitates the transfer of semantic knowledge using pre-defined semantic descriptors, while “discriminative” is crucial for accurate category discrimination. However, these two concepts are arguably inherently conflicting, as semantic descriptors are not specifically designed for image classification. Existing methods often struggle with balancing these two aspects and neglect the conflict between them, leading to suboptimal representation generalization and transferability to unseen classes. To address this issue, we propose a novel partially-shared multi-task representation learning method, termed PS-GZSL, which jointly preserves complementary and sharable knowledge between these two concepts. Specifically, we first propose a novel perspective that treats the learning of discriminative and semantic-relevant representations as optimizing a discrimination task and a visual-semantic alignment task, respectively. Then, to learn more complete and generalizable representations, PS-GZSL explicitly factorizes visual features into task-shared and task-specific representations and introduces two advanced tasks: an instance-level contrastive discrimination task and a relation-based visual-semantic alignment task. Furthermore, PS-GZSL employs Mixture-of-Experts (MoE) with a dropout mechanism to prevent representation degeneration and integrates a conditional GAN (cGAN) to synthesize unseen features for estimating unseen visual features. Extensive experiments and more competitive results on five widely-used GZSL benchmark datasets validate the effectiveness of our PS-GZSL.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/electronics12092085</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2079-9292 |
ispartof | Electronics (Basel), 2023-05, Vol.12 (9), p.2085 |
issn | 2079-9292 2079-9292 |
language | eng |
recordid | cdi_proquest_journals_2812386848 |
source | MDPI - Multidisciplinary Digital Publishing Institute; EZB-FREE-00999 freely available EZB journals |
subjects | Alignment Automatic classification Classification Degeneration Image classification Image processing Knowledge management Learning Machine learning Methods Neural networks Representations Semantics Visual discrimination Visual tasks Zero-shot learning |
title | Generalized Zero-Shot Image Classification via Partially-Shared Multi-Task Representation Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T19%3A40%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Generalized%20Zero-Shot%20Image%20Classification%20via%20Partially-Shared%20Multi-Task%20Representation%20Learning&rft.jtitle=Electronics%20(Basel)&rft.au=Wang,%20Gerui&rft.date=2023-05-03&rft.volume=12&rft.issue=9&rft.spage=2085&rft.pages=2085-&rft.issn=2079-9292&rft.eissn=2079-9292&rft_id=info:doi/10.3390/electronics12092085&rft_dat=%3Cgale_proqu%3EA749097176%3C/gale_proqu%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2812386848&rft_id=info:pmid/&rft_galeid=A749097176&rfr_iscdi=true |