A bi-network architecture for occlusion handling in Person re-identification
Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few...
Gespeichert in:
Veröffentlicht in: | Signal, image and video processing image and video processing, 2022-06, Vol.16 (4), p.1071-1079 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1079 |
---|---|
container_issue | 4 |
container_start_page | 1071 |
container_title | Signal, image and video processing |
container_volume | 16 |
creator | Tagore, Nirbhay Kumar Chattopadhyay, Pratik |
description | Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as
Occlusion Handling GAN
(
OHGAN
), is based on the popular U-Net architecture and is trained using
L2
loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our
OHGAN
performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using
Part-based Convolution Baseline (PCB)
. We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques. |
doi_str_mv | 10.1007/s11760-021-02056-4 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2651149375</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2651149375</sourcerecordid><originalsourceid>FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</originalsourceid><addsrcrecordid>eNp9kE9LAzEQxYMoWGq_gKcFz9FMsslujqX4Dxb0oOeQZpM2tWZrkkX89kZX9ObAMMPwe2_gIXQO5BIIaa4SQCMIJhRKEy5wfYRm0AqGoQE4_t0JO0WLlHakFKNNK9oZ6pbV2uNg8_sQXyodzdZna_IYbeWGWA3G7Mfkh1Btdej3PmwqH6pHG1M5RYt9b0P2zhudC3SGTpzeJ7v4mXP0fHP9tLrD3cPt_WrZYUNrmTGVRAjZa-LWEjiwlrM1E7VhTEpHjDXCOqMF131LKbRaS2kbrallhEtghs3RxeR7iMPbaFNWu2GMobxUVHCAWrKGF4pOlIlDStE6dYj-VccPBUR9Baem4FQJTn0Hp-oiYpMoFThsbPyz_kf1Cdl7b_o</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2651149375</pqid></control><display><type>article</type><title>A bi-network architecture for occlusion handling in Person re-identification</title><source>SpringerLink Journals - AutoHoldings</source><creator>Tagore, Nirbhay Kumar ; Chattopadhyay, Pratik</creator><creatorcontrib>Tagore, Nirbhay Kumar ; Chattopadhyay, Pratik</creatorcontrib><description>Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as
Occlusion Handling GAN
(
OHGAN
), is based on the popular U-Net architecture and is trained using
L2
loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our
OHGAN
performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using
Part-based Convolution Baseline (PCB)
. We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.</description><identifier>ISSN: 1863-1703</identifier><identifier>EISSN: 1863-1711</identifier><identifier>DOI: 10.1007/s11760-021-02056-4</identifier><language>eng</language><publisher>London: Springer London</publisher><subject>Cameras ; Computer architecture ; Computer Imaging ; Computer Science ; Datasets ; Handling ; Identification ; Image Processing and Computer Vision ; Multimedia Information Systems ; Occlusion ; Original Paper ; Pattern Recognition and Graphics ; Reconstruction ; Signal,Image and Speech Processing ; Vision</subject><ispartof>Signal, image and video processing, 2022-06, Vol.16 (4), p.1071-1079</ispartof><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021</rights><rights>The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2021.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</citedby><cites>FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</cites><orcidid>0000-0002-5805-6563</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11760-021-02056-4$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11760-021-02056-4$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27915,27916,41479,42548,51310</link.rule.ids></links><search><creatorcontrib>Tagore, Nirbhay Kumar</creatorcontrib><creatorcontrib>Chattopadhyay, Pratik</creatorcontrib><title>A bi-network architecture for occlusion handling in Person re-identification</title><title>Signal, image and video processing</title><addtitle>SIViP</addtitle><description>Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as
Occlusion Handling GAN
(
OHGAN
), is based on the popular U-Net architecture and is trained using
L2
loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our
OHGAN
performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using
Part-based Convolution Baseline (PCB)
. We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.</description><subject>Cameras</subject><subject>Computer architecture</subject><subject>Computer Imaging</subject><subject>Computer Science</subject><subject>Datasets</subject><subject>Handling</subject><subject>Identification</subject><subject>Image Processing and Computer Vision</subject><subject>Multimedia Information Systems</subject><subject>Occlusion</subject><subject>Original Paper</subject><subject>Pattern Recognition and Graphics</subject><subject>Reconstruction</subject><subject>Signal,Image and Speech Processing</subject><subject>Vision</subject><issn>1863-1703</issn><issn>1863-1711</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kE9LAzEQxYMoWGq_gKcFz9FMsslujqX4Dxb0oOeQZpM2tWZrkkX89kZX9ObAMMPwe2_gIXQO5BIIaa4SQCMIJhRKEy5wfYRm0AqGoQE4_t0JO0WLlHakFKNNK9oZ6pbV2uNg8_sQXyodzdZna_IYbeWGWA3G7Mfkh1Btdej3PmwqH6pHG1M5RYt9b0P2zhudC3SGTpzeJ7v4mXP0fHP9tLrD3cPt_WrZYUNrmTGVRAjZa-LWEjiwlrM1E7VhTEpHjDXCOqMF131LKbRaS2kbrallhEtghs3RxeR7iMPbaFNWu2GMobxUVHCAWrKGF4pOlIlDStE6dYj-VccPBUR9Baem4FQJTn0Hp-oiYpMoFThsbPyz_kf1Cdl7b_o</recordid><startdate>20220601</startdate><enddate>20220601</enddate><creator>Tagore, Nirbhay Kumar</creator><creator>Chattopadhyay, Pratik</creator><general>Springer London</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0000-0002-5805-6563</orcidid></search><sort><creationdate>20220601</creationdate><title>A bi-network architecture for occlusion handling in Person re-identification</title><author>Tagore, Nirbhay Kumar ; Chattopadhyay, Pratik</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c249t-290669da0fb91513853b364c3399f0cec6efca65ad82218aa99e7aa2e305913c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Cameras</topic><topic>Computer architecture</topic><topic>Computer Imaging</topic><topic>Computer Science</topic><topic>Datasets</topic><topic>Handling</topic><topic>Identification</topic><topic>Image Processing and Computer Vision</topic><topic>Multimedia Information Systems</topic><topic>Occlusion</topic><topic>Original Paper</topic><topic>Pattern Recognition and Graphics</topic><topic>Reconstruction</topic><topic>Signal,Image and Speech Processing</topic><topic>Vision</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Tagore, Nirbhay Kumar</creatorcontrib><creatorcontrib>Chattopadhyay, Pratik</creatorcontrib><collection>CrossRef</collection><jtitle>Signal, image and video processing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tagore, Nirbhay Kumar</au><au>Chattopadhyay, Pratik</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A bi-network architecture for occlusion handling in Person re-identification</atitle><jtitle>Signal, image and video processing</jtitle><stitle>SIViP</stitle><date>2022-06-01</date><risdate>2022</risdate><volume>16</volume><issue>4</issue><spage>1071</spage><epage>1079</epage><pages>1071-1079</pages><issn>1863-1703</issn><eissn>1863-1711</eissn><abstract>Person re-identification in a multi-camera setup is very important for tracking and monitoring the movement of individuals in public places. It is not always possible to capture human shape accurately using surveillance cameras due to occlusion caused by other individuals and/or objects. Only a few existing approaches consider the challenging problem of occlusion handling in person re-identification. We propose an effective bi-network architecture to carry out re-identification after occlusion reconstruction. Our architecture, termed as
Occlusion Handling GAN
(
OHGAN
), is based on the popular U-Net architecture and is trained using
L2
loss and binary cross-entropy loss. Due to unavailability of re-identification datasets with occlusion, the gallery set to train the network has been generated by synthetically adding occlusion of varying degrees to existing non-occluded datasets. Qualitative results show that our
OHGAN
performs reconstruction of occluded frames quite satisfactorily. Next, re-identification using the reconstructed frames has been performed using
Part-based Convolution Baseline (PCB)
. We carry out extensive experiments and compare the results of our proposed method with 11 state-of-the-art approaches on four public datasets. Results show that our method outperforms all other existing techniques.</abstract><cop>London</cop><pub>Springer London</pub><doi>10.1007/s11760-021-02056-4</doi><tpages>9</tpages><orcidid>https://orcid.org/0000-0002-5805-6563</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1863-1703 |
ispartof | Signal, image and video processing, 2022-06, Vol.16 (4), p.1071-1079 |
issn | 1863-1703 1863-1711 |
language | eng |
recordid | cdi_proquest_journals_2651149375 |
source | SpringerLink Journals - AutoHoldings |
subjects | Cameras Computer architecture Computer Imaging Computer Science Datasets Handling Identification Image Processing and Computer Vision Multimedia Information Systems Occlusion Original Paper Pattern Recognition and Graphics Reconstruction Signal,Image and Speech Processing Vision |
title | A bi-network architecture for occlusion handling in Person re-identification |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T23%3A51%3A43IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20bi-network%20architecture%20for%20occlusion%20handling%20in%20Person%20re-identification&rft.jtitle=Signal,%20image%20and%20video%20processing&rft.au=Tagore,%20Nirbhay%20Kumar&rft.date=2022-06-01&rft.volume=16&rft.issue=4&rft.spage=1071&rft.epage=1079&rft.pages=1071-1079&rft.issn=1863-1703&rft.eissn=1863-1711&rft_id=info:doi/10.1007/s11760-021-02056-4&rft_dat=%3Cproquest_cross%3E2651149375%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2651149375&rft_id=info:pmid/&rfr_iscdi=true |