Independent metric learning with aligned multi-part features for video-based person re-identification
Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recogniti...
Gespeichert in:
Veröffentlicht in: | Multimedia tools and applications 2019-10, Vol.78 (20), p.29323-29341 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 29341 |
---|---|
container_issue | 20 |
container_start_page | 29323 |
container_title | Multimedia tools and applications |
container_volume | 78 |
creator | Wu, Jingjing Jiang, Jianguo Qi, Meibin Liu, Hao |
description | Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method. |
doi_str_mv | 10.1007/s11042-018-7119-6 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2162912856</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2162912856</sourcerecordid><originalsourceid>FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</originalsourceid><addsrcrecordid>eNp1kE1LxDAQhosouK7-AG8Bz9FM2iTtURY_Fha86Dmk7WTN0k1rkir-e1sqePIyM4fnfQeeLLsGdguMqbsIwApOGZRUAVRUnmQrECqnSnE4ne68ZFQJBufZRYwHxkAKXqwy3PoWB5yGT-SIKbiGdGiCd35Pvlx6J6Zze48tOY5dcnQwIRGLJo0BI7F9IJ-uxZ7WJk7MgCH2ngSkbi501jUmud5fZmfWdBGvfvc6e3t8eN08093L03Zzv6NNDjLR2tZtoyqQFRcKm9KWhplclFXdKpQsr3MuyoIrbhTPlZzgAnnVikI0ZVtbka-zm6V3CP3HiDHpQz8GP73UHCSvgJdCThQsVBP6GANaPQR3NOFbA9OzTb3Y1JNNPdvUc4YvmTixfo_hr_n_0A8O4Xjh</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2162912856</pqid></control><display><type>article</type><title>Independent metric learning with aligned multi-part features for video-based person re-identification</title><source>SpringerLink (Online service)</source><creator>Wu, Jingjing ; Jiang, Jianguo ; Qi, Meibin ; Liu, Hao</creator><creatorcontrib>Wu, Jingjing ; Jiang, Jianguo ; Qi, Meibin ; Liu, Hao</creatorcontrib><description>Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-018-7119-6</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Alignment ; Cameras ; Computer Communication Networks ; Computer Science ; Data Structures and Information Theory ; Datasets ; Experiments ; Feature extraction ; Identification ; Kernels ; Learning ; Methods ; Multimedia ; Multimedia Information Systems ; Neural networks ; Occlusion ; Pedestrians ; Special Purpose and Application-Based Systems ; Surveillance</subject><ispartof>Multimedia tools and applications, 2019-10, Vol.78 (20), p.29323-29341</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2019</rights><rights>Multimedia Tools and Applications is a copyright of Springer, (2019). All Rights Reserved.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</citedby><cites>FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</cites><orcidid>0000-0002-3818-4277</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-018-7119-6$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-018-7119-6$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>315,781,785,27926,27927,41490,42559,51321</link.rule.ids></links><search><creatorcontrib>Wu, Jingjing</creatorcontrib><creatorcontrib>Jiang, Jianguo</creatorcontrib><creatorcontrib>Qi, Meibin</creatorcontrib><creatorcontrib>Liu, Hao</creatorcontrib><title>Independent metric learning with aligned multi-part features for video-based person re-identification</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.</description><subject>Alignment</subject><subject>Cameras</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Datasets</subject><subject>Experiments</subject><subject>Feature extraction</subject><subject>Identification</subject><subject>Kernels</subject><subject>Learning</subject><subject>Methods</subject><subject>Multimedia</subject><subject>Multimedia Information Systems</subject><subject>Neural networks</subject><subject>Occlusion</subject><subject>Pedestrians</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Surveillance</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp1kE1LxDAQhosouK7-AG8Bz9FM2iTtURY_Fha86Dmk7WTN0k1rkir-e1sqePIyM4fnfQeeLLsGdguMqbsIwApOGZRUAVRUnmQrECqnSnE4ne68ZFQJBufZRYwHxkAKXqwy3PoWB5yGT-SIKbiGdGiCd35Pvlx6J6Zze48tOY5dcnQwIRGLJo0BI7F9IJ-uxZ7WJk7MgCH2ngSkbi501jUmud5fZmfWdBGvfvc6e3t8eN08093L03Zzv6NNDjLR2tZtoyqQFRcKm9KWhplclFXdKpQsr3MuyoIrbhTPlZzgAnnVikI0ZVtbka-zm6V3CP3HiDHpQz8GP73UHCSvgJdCThQsVBP6GANaPQR3NOFbA9OzTb3Y1JNNPdvUc4YvmTixfo_hr_n_0A8O4Xjh</recordid><startdate>20191001</startdate><enddate>20191001</enddate><creator>Wu, Jingjing</creator><creator>Jiang, Jianguo</creator><creator>Qi, Meibin</creator><creator>Liu, Hao</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-3818-4277</orcidid></search><sort><creationdate>20191001</creationdate><title>Independent metric learning with aligned multi-part features for video-based person re-identification</title><author>Wu, Jingjing ; Jiang, Jianguo ; Qi, Meibin ; Liu, Hao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Alignment</topic><topic>Cameras</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Datasets</topic><topic>Experiments</topic><topic>Feature extraction</topic><topic>Identification</topic><topic>Kernels</topic><topic>Learning</topic><topic>Methods</topic><topic>Multimedia</topic><topic>Multimedia Information Systems</topic><topic>Neural networks</topic><topic>Occlusion</topic><topic>Pedestrians</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Surveillance</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wu, Jingjing</creatorcontrib><creatorcontrib>Jiang, Jianguo</creatorcontrib><creatorcontrib>Qi, Meibin</creatorcontrib><creatorcontrib>Liu, Hao</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Access via ABI/INFORM (ProQuest)</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Database (1962 - current)</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer science database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>ProQuest research library</collection><collection>Research Library (Corporate)</collection><collection>ProQuest advanced technologies & aerospace journals</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wu, Jingjing</au><au>Jiang, Jianguo</au><au>Qi, Meibin</au><au>Liu, Hao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Independent metric learning with aligned multi-part features for video-based person re-identification</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2019-10-01</date><risdate>2019</risdate><volume>78</volume><issue>20</issue><spage>29323</spage><epage>29341</epage><pages>29323-29341</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-018-7119-6</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-3818-4277</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1380-7501 |
ispartof | Multimedia tools and applications, 2019-10, Vol.78 (20), p.29323-29341 |
issn | 1380-7501 1573-7721 |
language | eng |
recordid | cdi_proquest_journals_2162912856 |
source | SpringerLink (Online service) |
subjects | Alignment Cameras Computer Communication Networks Computer Science Data Structures and Information Theory Datasets Experiments Feature extraction Identification Kernels Learning Methods Multimedia Multimedia Information Systems Neural networks Occlusion Pedestrians Special Purpose and Application-Based Systems Surveillance |
title | Independent metric learning with aligned multi-part features for video-based person re-identification |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T22%3A17%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Independent%20metric%20learning%20with%20aligned%20multi-part%20features%20for%20video-based%20person%20re-identification&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Wu,%20Jingjing&rft.date=2019-10-01&rft.volume=78&rft.issue=20&rft.spage=29323&rft.epage=29341&rft.pages=29323-29341&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-018-7119-6&rft_dat=%3Cproquest_cross%3E2162912856%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2162912856&rft_id=info:pmid/&rfr_iscdi=true |