PAFormer: Part Aware Transformer for Person Re-identification

Within the domain of person re-identification (ReID), partial ReID methods are considered mainstream, aiming to measure feature distances through comparisons of body parts between samples. However, in practice, previous methods often lack sufficient awareness of anatomical aspect of body parts, resu...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Jung, Hyeono, Lee, Jangwon, Yoo, Jiwon, Ko, Dami, Kim, Gyeonghwan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Jung, Hyeono
Lee, Jangwon
Yoo, Jiwon
Ko, Dami
Kim, Gyeonghwan
description Within the domain of person re-identification (ReID), partial ReID methods are considered mainstream, aiming to measure feature distances through comparisons of body parts between samples. However, in practice, previous methods often lack sufficient awareness of anatomical aspect of body parts, resulting in the failure to capture features of the same body parts across different samples. To address this issue, we introduce \textbf{Part Aware Transformer (PAFormer)}, a pose estimation based ReID model which can perform precise part-to-part comparison. In order to inject part awareness to pose tokens, we introduce learnable parameters called `pose token' which estimate the correlation between each body part and partial regions of the image. Notably, at inference phase, PAFormer operates without additional modules related to body part localization, which is commonly used in previous ReID methodologies leveraging pose estimation models. Additionally, leveraging the enhanced awareness of body parts, PAFormer suggests the use of a learning-based visibility predictor to estimate the degree of occlusion for each body part. Also, we introduce a teacher forcing technique using ground truth visibility scores which enables PAFormer to be trained only with visible parts. A set of extensive experiments show that our method outperforms existing approaches on well-known ReID benchmark datasets.
doi_str_mv 10.48550/arxiv.2408.05918
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2408_05918</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2408_05918</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2408_059183</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjGw0DMwtTS04GSwDXB0yy_KTS2yUghILCpRcCxPLEpVCClKzCtOA4srACmFgNSi4vw8haBU3cyU1LySzLTM5MSSzPw8HgbWtMSc4lReKM3NIO_mGuLsoQu2KL6gKDM3sagyHmRhPNhCY8IqAAwWNVw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>PAFormer: Part Aware Transformer for Person Re-identification</title><source>arXiv.org</source><creator>Jung, Hyeono ; Lee, Jangwon ; Yoo, Jiwon ; Ko, Dami ; Kim, Gyeonghwan</creator><creatorcontrib>Jung, Hyeono ; Lee, Jangwon ; Yoo, Jiwon ; Ko, Dami ; Kim, Gyeonghwan</creatorcontrib><description>Within the domain of person re-identification (ReID), partial ReID methods are considered mainstream, aiming to measure feature distances through comparisons of body parts between samples. However, in practice, previous methods often lack sufficient awareness of anatomical aspect of body parts, resulting in the failure to capture features of the same body parts across different samples. To address this issue, we introduce \textbf{Part Aware Transformer (PAFormer)}, a pose estimation based ReID model which can perform precise part-to-part comparison. In order to inject part awareness to pose tokens, we introduce learnable parameters called `pose token' which estimate the correlation between each body part and partial regions of the image. Notably, at inference phase, PAFormer operates without additional modules related to body part localization, which is commonly used in previous ReID methodologies leveraging pose estimation models. Additionally, leveraging the enhanced awareness of body parts, PAFormer suggests the use of a learning-based visibility predictor to estimate the degree of occlusion for each body part. Also, we introduce a teacher forcing technique using ground truth visibility scores which enables PAFormer to be trained only with visible parts. A set of extensive experiments show that our method outperforms existing approaches on well-known ReID benchmark datasets.</description><identifier>DOI: 10.48550/arxiv.2408.05918</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2024-08</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2408.05918$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2408.05918$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Jung, Hyeono</creatorcontrib><creatorcontrib>Lee, Jangwon</creatorcontrib><creatorcontrib>Yoo, Jiwon</creatorcontrib><creatorcontrib>Ko, Dami</creatorcontrib><creatorcontrib>Kim, Gyeonghwan</creatorcontrib><title>PAFormer: Part Aware Transformer for Person Re-identification</title><description>Within the domain of person re-identification (ReID), partial ReID methods are considered mainstream, aiming to measure feature distances through comparisons of body parts between samples. However, in practice, previous methods often lack sufficient awareness of anatomical aspect of body parts, resulting in the failure to capture features of the same body parts across different samples. To address this issue, we introduce \textbf{Part Aware Transformer (PAFormer)}, a pose estimation based ReID model which can perform precise part-to-part comparison. In order to inject part awareness to pose tokens, we introduce learnable parameters called `pose token' which estimate the correlation between each body part and partial regions of the image. Notably, at inference phase, PAFormer operates without additional modules related to body part localization, which is commonly used in previous ReID methodologies leveraging pose estimation models. Additionally, leveraging the enhanced awareness of body parts, PAFormer suggests the use of a learning-based visibility predictor to estimate the degree of occlusion for each body part. Also, we introduce a teacher forcing technique using ground truth visibility scores which enables PAFormer to be trained only with visible parts. A set of extensive experiments show that our method outperforms existing approaches on well-known ReID benchmark datasets.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjGw0DMwtTS04GSwDXB0yy_KTS2yUghILCpRcCxPLEpVCClKzCtOA4srACmFgNSi4vw8haBU3cyU1LySzLTM5MSSzPw8HgbWtMSc4lReKM3NIO_mGuLsoQu2KL6gKDM3sagyHmRhPNhCY8IqAAwWNVw</recordid><startdate>20240812</startdate><enddate>20240812</enddate><creator>Jung, Hyeono</creator><creator>Lee, Jangwon</creator><creator>Yoo, Jiwon</creator><creator>Ko, Dami</creator><creator>Kim, Gyeonghwan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240812</creationdate><title>PAFormer: Part Aware Transformer for Person Re-identification</title><author>Jung, Hyeono ; Lee, Jangwon ; Yoo, Jiwon ; Ko, Dami ; Kim, Gyeonghwan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2408_059183</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Jung, Hyeono</creatorcontrib><creatorcontrib>Lee, Jangwon</creatorcontrib><creatorcontrib>Yoo, Jiwon</creatorcontrib><creatorcontrib>Ko, Dami</creatorcontrib><creatorcontrib>Kim, Gyeonghwan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Jung, Hyeono</au><au>Lee, Jangwon</au><au>Yoo, Jiwon</au><au>Ko, Dami</au><au>Kim, Gyeonghwan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>PAFormer: Part Aware Transformer for Person Re-identification</atitle><date>2024-08-12</date><risdate>2024</risdate><abstract>Within the domain of person re-identification (ReID), partial ReID methods are considered mainstream, aiming to measure feature distances through comparisons of body parts between samples. However, in practice, previous methods often lack sufficient awareness of anatomical aspect of body parts, resulting in the failure to capture features of the same body parts across different samples. To address this issue, we introduce \textbf{Part Aware Transformer (PAFormer)}, a pose estimation based ReID model which can perform precise part-to-part comparison. In order to inject part awareness to pose tokens, we introduce learnable parameters called `pose token' which estimate the correlation between each body part and partial regions of the image. Notably, at inference phase, PAFormer operates without additional modules related to body part localization, which is commonly used in previous ReID methodologies leveraging pose estimation models. Additionally, leveraging the enhanced awareness of body parts, PAFormer suggests the use of a learning-based visibility predictor to estimate the degree of occlusion for each body part. Also, we introduce a teacher forcing technique using ground truth visibility scores which enables PAFormer to be trained only with visible parts. A set of extensive experiments show that our method outperforms existing approaches on well-known ReID benchmark datasets.</abstract><doi>10.48550/arxiv.2408.05918</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2408.05918
ispartof
issn
language eng
recordid cdi_arxiv_primary_2408_05918
source arXiv.org
subjects Computer Science - Computer Vision and Pattern Recognition
title PAFormer: Part Aware Transformer for Person Re-identification
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T11%3A51%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=PAFormer:%20Part%20Aware%20Transformer%20for%20Person%20Re-identification&rft.au=Jung,%20Hyeono&rft.date=2024-08-12&rft_id=info:doi/10.48550/arxiv.2408.05918&rft_dat=%3Carxiv_GOX%3E2408_05918%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true