Independent metric learning with aligned multi-part features for video-based person re-identification

Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recogniti...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Multimedia tools and applications 2019-10, Vol.78 (20), p.29323-29341
Hauptverfasser: Wu, Jingjing, Jiang, Jianguo, Qi, Meibin, Liu, Hao
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 29341
container_issue 20
container_start_page 29323
container_title Multimedia tools and applications
container_volume 78
creator Wu, Jingjing
Jiang, Jianguo
Qi, Meibin
Liu, Hao
description Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.
doi_str_mv 10.1007/s11042-018-7119-6
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2162912856</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2162912856</sourcerecordid><originalsourceid>FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</originalsourceid><addsrcrecordid>eNp1kE1LxDAQhosouK7-AG8Bz9FM2iTtURY_Fha86Dmk7WTN0k1rkir-e1sqePIyM4fnfQeeLLsGdguMqbsIwApOGZRUAVRUnmQrECqnSnE4ne68ZFQJBufZRYwHxkAKXqwy3PoWB5yGT-SIKbiGdGiCd35Pvlx6J6Zze48tOY5dcnQwIRGLJo0BI7F9IJ-uxZ7WJk7MgCH2ngSkbi501jUmud5fZmfWdBGvfvc6e3t8eN08093L03Zzv6NNDjLR2tZtoyqQFRcKm9KWhplclFXdKpQsr3MuyoIrbhTPlZzgAnnVikI0ZVtbka-zm6V3CP3HiDHpQz8GP73UHCSvgJdCThQsVBP6GANaPQR3NOFbA9OzTb3Y1JNNPdvUc4YvmTixfo_hr_n_0A8O4Xjh</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2162912856</pqid></control><display><type>article</type><title>Independent metric learning with aligned multi-part features for video-based person re-identification</title><source>SpringerLink (Online service)</source><creator>Wu, Jingjing ; Jiang, Jianguo ; Qi, Meibin ; Liu, Hao</creator><creatorcontrib>Wu, Jingjing ; Jiang, Jianguo ; Qi, Meibin ; Liu, Hao</creatorcontrib><description>Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-018-7119-6</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Alignment ; Cameras ; Computer Communication Networks ; Computer Science ; Data Structures and Information Theory ; Datasets ; Experiments ; Feature extraction ; Identification ; Kernels ; Learning ; Methods ; Multimedia ; Multimedia Information Systems ; Neural networks ; Occlusion ; Pedestrians ; Special Purpose and Application-Based Systems ; Surveillance</subject><ispartof>Multimedia tools and applications, 2019-10, Vol.78 (20), p.29323-29341</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2019</rights><rights>Multimedia Tools and Applications is a copyright of Springer, (2019). All Rights Reserved.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</citedby><cites>FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</cites><orcidid>0000-0002-3818-4277</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-018-7119-6$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-018-7119-6$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>315,781,785,27926,27927,41490,42559,51321</link.rule.ids></links><search><creatorcontrib>Wu, Jingjing</creatorcontrib><creatorcontrib>Jiang, Jianguo</creatorcontrib><creatorcontrib>Qi, Meibin</creatorcontrib><creatorcontrib>Liu, Hao</creatorcontrib><title>Independent metric learning with aligned multi-part features for video-based person re-identification</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.</description><subject>Alignment</subject><subject>Cameras</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Datasets</subject><subject>Experiments</subject><subject>Feature extraction</subject><subject>Identification</subject><subject>Kernels</subject><subject>Learning</subject><subject>Methods</subject><subject>Multimedia</subject><subject>Multimedia Information Systems</subject><subject>Neural networks</subject><subject>Occlusion</subject><subject>Pedestrians</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Surveillance</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp1kE1LxDAQhosouK7-AG8Bz9FM2iTtURY_Fha86Dmk7WTN0k1rkir-e1sqePIyM4fnfQeeLLsGdguMqbsIwApOGZRUAVRUnmQrECqnSnE4ne68ZFQJBufZRYwHxkAKXqwy3PoWB5yGT-SIKbiGdGiCd35Pvlx6J6Zze48tOY5dcnQwIRGLJo0BI7F9IJ-uxZ7WJk7MgCH2ngSkbi501jUmud5fZmfWdBGvfvc6e3t8eN08093L03Zzv6NNDjLR2tZtoyqQFRcKm9KWhplclFXdKpQsr3MuyoIrbhTPlZzgAnnVikI0ZVtbka-zm6V3CP3HiDHpQz8GP73UHCSvgJdCThQsVBP6GANaPQR3NOFbA9OzTb3Y1JNNPdvUc4YvmTixfo_hr_n_0A8O4Xjh</recordid><startdate>20191001</startdate><enddate>20191001</enddate><creator>Wu, Jingjing</creator><creator>Jiang, Jianguo</creator><creator>Qi, Meibin</creator><creator>Liu, Hao</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-3818-4277</orcidid></search><sort><creationdate>20191001</creationdate><title>Independent metric learning with aligned multi-part features for video-based person re-identification</title><author>Wu, Jingjing ; Jiang, Jianguo ; Qi, Meibin ; Liu, Hao</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c316t-bfbdc79169257ec8f8a0a3589bd7e603b32584272a72376bdc4e29d545c8dbf53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Alignment</topic><topic>Cameras</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Datasets</topic><topic>Experiments</topic><topic>Feature extraction</topic><topic>Identification</topic><topic>Kernels</topic><topic>Learning</topic><topic>Methods</topic><topic>Multimedia</topic><topic>Multimedia Information Systems</topic><topic>Neural networks</topic><topic>Occlusion</topic><topic>Pedestrians</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Surveillance</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wu, Jingjing</creatorcontrib><creatorcontrib>Jiang, Jianguo</creatorcontrib><creatorcontrib>Qi, Meibin</creatorcontrib><creatorcontrib>Liu, Hao</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Access via ABI/INFORM (ProQuest)</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Database‎ (1962 - current)</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer science database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>ProQuest research library</collection><collection>Research Library (Corporate)</collection><collection>ProQuest advanced technologies &amp; aerospace journals</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wu, Jingjing</au><au>Jiang, Jianguo</au><au>Qi, Meibin</au><au>Liu, Hao</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Independent metric learning with aligned multi-part features for video-based person re-identification</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2019-10-01</date><risdate>2019</risdate><volume>78</volume><issue>20</issue><spage>29323</spage><epage>29341</epage><pages>29323-29341</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>Video-based person re-identification attracts wide attention because it plays a crucial role for many applications in the video surveillance. The task of video-based person re-identification is to match image sequences of the pedestrian recorded by non-overlapping cameras. Like many visual recognition problems, variations in pose, viewpoints, illumination, and occlusion make this task non-trivial. Aiming at increasing the robustness of features to variations and occlusion, this paper designs an aligned multi-part image model inspired by human visual attention mechanism. This model performs a pose estimation method to align the pedestrians. Then, it divides the images to extract multi-part appearance features. Besides, we present independent metric learning to combine the multi-part appearance and spatial-temporal features, which obtains several metric kernels by feeding these features into distance metric learning respectively. These kernels are fused with the weights learned by the attention measure. The novel way of features fusion can achieve better functional complementarity of these features. In experiments, we analyze the effectiveness of the major components. Extensive experiments on two public benchmark datasets, i.e., the iLIDS-VID and PRID-2011 datasets, demonstrate the effectiveness of the proposed method.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-018-7119-6</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-3818-4277</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1380-7501
ispartof Multimedia tools and applications, 2019-10, Vol.78 (20), p.29323-29341
issn 1380-7501
1573-7721
language eng
recordid cdi_proquest_journals_2162912856
source SpringerLink (Online service)
subjects Alignment
Cameras
Computer Communication Networks
Computer Science
Data Structures and Information Theory
Datasets
Experiments
Feature extraction
Identification
Kernels
Learning
Methods
Multimedia
Multimedia Information Systems
Neural networks
Occlusion
Pedestrians
Special Purpose and Application-Based Systems
Surveillance
title Independent metric learning with aligned multi-part features for video-based person re-identification
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T22%3A17%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Independent%20metric%20learning%20with%20aligned%20multi-part%20features%20for%20video-based%20person%20re-identification&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Wu,%20Jingjing&rft.date=2019-10-01&rft.volume=78&rft.issue=20&rft.spage=29323&rft.epage=29341&rft.pages=29323-29341&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-018-7119-6&rft_dat=%3Cproquest_cross%3E2162912856%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2162912856&rft_id=info:pmid/&rfr_iscdi=true