A bi-network architecture for occlusion handling in Person re-identification

Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Signal, image and video processing image and video processing, 2022-06, Vol.16 (4), p.1071-1079
Hauptverfasser: Tagore, Nirbhay Kumar, Chattopadhyay, Pratik
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1079
container_issue 4
container_start_page 1071
container_title Signal, image and video processing
container_volume 16
creator Tagore, Nirbhay Kumar
Chattopadhyay, Pratik
description Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as Occlusion Handling GAN ( OHGAN ), is based on the popular U-Net architecture and is trained using L2 loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our OHGAN performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using Part-based Convolution Baseline (PCB) . We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.
doi_str_mv 10.1007/s11760-021-02056-4
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2651149375</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2651149375</sourcerecordid><originalsourceid>FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</originalsourceid><addsrcrecordid>eNp9kE9LAzEQxYMoWGq_gKcFz9FMsslujqX4Dxb0oOeQZpM2tWZrkkX89kZX9ObAMMPwe2_gIXQO5BIIaa4SQCMIJhRKEy5wfYRm0AqGoQE4_t0JO0WLlHakFKNNK9oZ6pbV2uNg8_sQXyodzdZna_IYbeWGWA3G7Mfkh1Btdej3PmwqH6pHG1M5RYt9b0P2zhudC3SGTpzeJ7v4mXP0fHP9tLrD3cPt_WrZYUNrmTGVRAjZa-LWEjiwlrM1E7VhTEpHjDXCOqMF131LKbRaS2kbrallhEtghs3RxeR7iMPbaFNWu2GMobxUVHCAWrKGF4pOlIlDStE6dYj-VccPBUR9Baem4FQJTn0Hp-oiYpMoFThsbPyz_kf1Cdl7b_o</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2651149375</pqid></control><display><type>article</type><title>A bi-network architecture for occlusion handling in Person re-identification</title><source>SpringerLink Journals - AutoHoldings</source><creator>Tagore, Nirbhay Kumar ; Chattopadhyay, Pratik</creator><creatorcontrib>Tagore, Nirbhay Kumar ; Chattopadhyay, Pratik</creatorcontrib><description>Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as Occlusion Handling GAN ( OHGAN ), is based on the popular U-Net architecture and is trained using L2 loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our OHGAN performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using Part-based Convolution Baseline (PCB) . We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.</description><identifier>ISSN: 1863-1703</identifier><identifier>EISSN: 1863-1711</identifier><identifier>DOI: 10.1007/s11760-021-02056-4</identifier><language>eng</language><publisher>London: Springer London</publisher><subject>Cameras ; Computer architecture ; Computer Imaging ; Computer Science ; Datasets ; Handling ; Identification ; Image Processing and Computer Vision ; Multimedia Information Systems ; Occlusion ; Original Paper ; Pattern Recognition and Graphics ; Reconstruction ; Signal,Image and Speech Processing ; Vision</subject><ispartof>Signal, image and video processing, 2022-06, Vol.16 (4), p.1071-1079</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021</rights><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</citedby><cites>FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</cites><orcidid>0000-0002-5805-6563</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11760-021-02056-4$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11760-021-02056-4$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27915,27916,41479,42548,51310</link.rule.ids></links><search><creatorcontrib>Tagore, Nirbhay Kumar</creatorcontrib><creatorcontrib>Chattopadhyay, Pratik</creatorcontrib><title>A bi-network architecture for occlusion handling in Person re-identification</title><title>Signal, image and video processing</title><addtitle>SIViP</addtitle><description>Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as Occlusion Handling GAN ( OHGAN ), is based on the popular U-Net architecture and is trained using L2 loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our OHGAN performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using Part-based Convolution Baseline (PCB) . We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.</description><subject>Cameras</subject><subject>Computer architecture</subject><subject>Computer Imaging</subject><subject>Computer Science</subject><subject>Datasets</subject><subject>Handling</subject><subject>Identification</subject><subject>Image Processing and Computer Vision</subject><subject>Multimedia Information Systems</subject><subject>Occlusion</subject><subject>Original Paper</subject><subject>Pattern Recognition and Graphics</subject><subject>Reconstruction</subject><subject>Signal,Image and Speech Processing</subject><subject>Vision</subject><issn>1863-1703</issn><issn>1863-1711</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kE9LAzEQxYMoWGq_gKcFz9FMsslujqX4Dxb0oOeQZpM2tWZrkkX89kZX9ObAMMPwe2_gIXQO5BIIaa4SQCMIJhRKEy5wfYRm0AqGoQE4_t0JO0WLlHakFKNNK9oZ6pbV2uNg8_sQXyodzdZna_IYbeWGWA3G7Mfkh1Btdej3PmwqH6pHG1M5RYt9b0P2zhudC3SGTpzeJ7v4mXP0fHP9tLrD3cPt_WrZYUNrmTGVRAjZa-LWEjiwlrM1E7VhTEpHjDXCOqMF131LKbRaS2kbrallhEtghs3RxeR7iMPbaFNWu2GMobxUVHCAWrKGF4pOlIlDStE6dYj-VccPBUR9Baem4FQJTn0Hp-oiYpMoFThsbPyz_kf1Cdl7b_o</recordid><startdate>20220601</startdate><enddate>20220601</enddate><creator>Tagore, Nirbhay Kumar</creator><creator>Chattopadhyay, Pratik</creator><general>Springer London</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-5805-6563</orcidid></search><sort><creationdate>20220601</creationdate><title>A bi-network architecture for occlusion handling in Person re-identification</title><author>Tagore, Nirbhay Kumar ; Chattopadhyay, Pratik</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Cameras</topic><topic>Computer architecture</topic><topic>Computer Imaging</topic><topic>Computer Science</topic><topic>Datasets</topic><topic>Handling</topic><topic>Identification</topic><topic>Image Processing and Computer Vision</topic><topic>Multimedia Information Systems</topic><topic>Occlusion</topic><topic>Original Paper</topic><topic>Pattern Recognition and Graphics</topic><topic>Reconstruction</topic><topic>Signal,Image and Speech Processing</topic><topic>Vision</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Tagore, Nirbhay Kumar</creatorcontrib><creatorcontrib>Chattopadhyay, Pratik</creatorcontrib><collection>CrossRef</collection><jtitle>Signal, image and video processing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tagore, Nirbhay Kumar</au><au>Chattopadhyay, Pratik</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A bi-network architecture for occlusion handling in Person re-identification</atitle><jtitle>Signal, image and video processing</jtitle><stitle>SIViP</stitle><date>2022-06-01</date><risdate>2022</risdate><volume>16</volume><issue>4</issue><spage>1071</spage><epage>1079</epage><pages>1071-1079</pages><issn>1863-1703</issn><eissn>1863-1711</eissn><abstract>Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as Occlusion Handling GAN ( OHGAN ), is based on the popular U-Net architecture and is trained using L2 loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our OHGAN performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using Part-based Convolution Baseline (PCB) . We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.</abstract><cop>London</cop><pub>Springer London</pub><doi>10.1007/s11760-021-02056-4</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0002-5805-6563</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1863-1703
ispartof Signal, image and video processing, 2022-06, Vol.16 (4), p.1071-1079
issn 1863-1703
1863-1711
language eng
recordid cdi_proquest_journals_2651149375
source SpringerLink Journals - AutoHoldings
subjects Cameras
Computer architecture
Computer Imaging
Computer Science
Datasets
Handling
Identification
Image Processing and Computer Vision
Multimedia Information Systems
Occlusion
Original Paper
Pattern Recognition and Graphics
Reconstruction
Signal,Image and Speech Processing
Vision
title A bi-network architecture for occlusion handling in Person re-identification
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T23%3A51%3A43IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20bi-network%20architecture%20for%20occlusion%20handling%20in%20Person%20re-identification&rft.jtitle=Signal,%20image%20and%20video%20processing&rft.au=Tagore,%20Nirbhay%20Kumar&rft.date=2022-06-01&rft.volume=16&rft.issue=4&rft.spage=1071&rft.epage=1079&rft.pages=1071-1079&rft.issn=1863-1703&rft.eissn=1863-1711&rft_id=info:doi/10.1007/s11760-021-02056-4&rft_dat=%3Cproquest_cross%3E2651149375%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2651149375&rft_id=info:pmid/&rfr_iscdi=true