Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval

Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on image processing 2017-03, Vol.26 (3), p.1393-1404
Hauptverfasser: Wang, Yang, Lin, Xuemin, Wu, Lin, Zhang, Wenjie
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1404
container_issue 3
container_start_page 1393
container_title IEEE transactions on image processing
container_volume 26
creator Wang, Yang
Lin, Xuemin
Wu, Lin
Zhang, Wenjie
description Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].
doi_str_mv 10.1109/TIP.2017.2655449
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_1861582580</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>7823000</ieee_id><sourcerecordid>1861582580</sourcerecordid><originalsourceid>FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</originalsourceid><addsrcrecordid>eNpdkMtP40AMh0erRQsU7iuthCJx4ZJizysTbqgUqFSeYk97iPJwpECaKTMJj_-eKS0c9mRL_mz9_DH2G2GMCOnxw-x2zAGTMddKSZn-YDuYSowBJP8ZelBJnKBMt9mu948AKBXqX2ybGwShlNlh_6Z1TWXfvFB0NbR9E98N5N6j6dsy73xjO38STWzb5oV1-Sd1RrSMrql_te7JR7V10b0tBt9H87yrFrl7iu6pdw295O0e26rz1tP-po7Y3_Ppw-Qynt9czCan87gUMunjRJTATa4JtUhVWSleFigKXXMsjdKJFkaqSinBiScK0xCdijrl3JjaqMqIETta3106-zyQ77NF40sKqTuyg8_QaFSGKwMBPfwPfbSD60K6FZVKDqB5oGBNlc5676jOlq4Jv71nCNlKfBbEZyvx2UZ8WDnYHB6KBVXfC1-mA_BnDTRE9D1ODBcAID4Ag6OFJw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1869420062</pqid></control><display><type>article</type><title>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</title><source>IEEE Electronic Library (IEL)</source><creator>Wang, Yang ; Lin, Xuemin ; Wu, Lin ; Zhang, Wenjie</creator><creatorcontrib>Wang, Yang ; Lin, Xuemin ; Wu, Lin ; Zhang, Wenjie</creatorcontrib><description>Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].</description><identifier>ISSN: 1057-7149</identifier><identifier>EISSN: 1941-0042</identifier><identifier>DOI: 10.1109/TIP.2017.2655449</identifier><identifier>PMID: 28103558</identifier><identifier>CODEN: IIPRE4</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Angles (geometry) ; Collaboration ; collaborative deep networks ; Digital media ; Dirichlet problem ; Filtration ; Geometry ; Landmark photo retrieval ; Landmarks ; Level (quantity) ; multi-query expansions ; Photography ; Poles and towers ; Ranking ; Retrieval ; Robustness ; Shape ; Social network services ; Social networks ; Visualization</subject><ispartof>IEEE transactions on image processing, 2017-03, Vol.26 (3), p.1393-1404</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2017</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</citedby><cites>FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</cites><orcidid>0000-0003-1029-9280</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/7823000$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27923,27924,54757</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/7823000$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/28103558$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Wang, Yang</creatorcontrib><creatorcontrib>Lin, Xuemin</creatorcontrib><creatorcontrib>Wu, Lin</creatorcontrib><creatorcontrib>Zhang, Wenjie</creatorcontrib><title>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</title><title>IEEE transactions on image processing</title><addtitle>TIP</addtitle><addtitle>IEEE Trans Image Process</addtitle><description>Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].</description><subject>Angles (geometry)</subject><subject>Collaboration</subject><subject>collaborative deep networks</subject><subject>Digital media</subject><subject>Dirichlet problem</subject><subject>Filtration</subject><subject>Geometry</subject><subject>Landmark photo retrieval</subject><subject>Landmarks</subject><subject>Level (quantity)</subject><subject>multi-query expansions</subject><subject>Photography</subject><subject>Poles and towers</subject><subject>Ranking</subject><subject>Retrieval</subject><subject>Robustness</subject><subject>Shape</subject><subject>Social network services</subject><subject>Social networks</subject><subject>Visualization</subject><issn>1057-7149</issn><issn>1941-0042</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2017</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpdkMtP40AMh0erRQsU7iuthCJx4ZJizysTbqgUqFSeYk97iPJwpECaKTMJj_-eKS0c9mRL_mz9_DH2G2GMCOnxw-x2zAGTMddKSZn-YDuYSowBJP8ZelBJnKBMt9mu948AKBXqX2ybGwShlNlh_6Z1TWXfvFB0NbR9E98N5N6j6dsy73xjO38STWzb5oV1-Sd1RrSMrql_te7JR7V10b0tBt9H87yrFrl7iu6pdw295O0e26rz1tP-po7Y3_Ppw-Qynt9czCan87gUMunjRJTATa4JtUhVWSleFigKXXMsjdKJFkaqSinBiScK0xCdijrl3JjaqMqIETta3106-zyQ77NF40sKqTuyg8_QaFSGKwMBPfwPfbSD60K6FZVKDqB5oGBNlc5676jOlq4Jv71nCNlKfBbEZyvx2UZ8WDnYHB6KBVXfC1-mA_BnDTRE9D1ODBcAID4Ag6OFJw</recordid><startdate>20170301</startdate><enddate>20170301</enddate><creator>Wang, Yang</creator><creator>Lin, Xuemin</creator><creator>Wu, Lin</creator><creator>Zhang, Wenjie</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-1029-9280</orcidid></search><sort><creationdate>20170301</creationdate><title>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</title><author>Wang, Yang ; Lin, Xuemin ; Wu, Lin ; Zhang, Wenjie</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2017</creationdate><topic>Angles (geometry)</topic><topic>Collaboration</topic><topic>collaborative deep networks</topic><topic>Digital media</topic><topic>Dirichlet problem</topic><topic>Filtration</topic><topic>Geometry</topic><topic>Landmark photo retrieval</topic><topic>Landmarks</topic><topic>Level (quantity)</topic><topic>multi-query expansions</topic><topic>Photography</topic><topic>Poles and towers</topic><topic>Ranking</topic><topic>Retrieval</topic><topic>Robustness</topic><topic>Shape</topic><topic>Social network services</topic><topic>Social networks</topic><topic>Visualization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Yang</creatorcontrib><creatorcontrib>Lin, Xuemin</creatorcontrib><creatorcontrib>Wu, Lin</creatorcontrib><creatorcontrib>Zhang, Wenjie</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transactions on image processing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wang, Yang</au><au>Lin, Xuemin</au><au>Wu, Lin</au><au>Zhang, Wenjie</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</atitle><jtitle>IEEE transactions on image processing</jtitle><stitle>TIP</stitle><addtitle>IEEE Trans Image Process</addtitle><date>2017-03-01</date><risdate>2017</risdate><volume>26</volume><issue>3</issue><spage>1393</spage><epage>1404</epage><pages>1393-1404</pages><issn>1057-7149</issn><eissn>1941-0042</eissn><coden>IIPRE4</coden><abstract>Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].</abstract><cop>United States</cop><pub>IEEE</pub><pmid>28103558</pmid><doi>10.1109/TIP.2017.2655449</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0003-1029-9280</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1057-7149
ispartof IEEE transactions on image processing, 2017-03, Vol.26 (3), p.1393-1404
issn 1057-7149
1941-0042
language eng
recordid cdi_proquest_miscellaneous_1861582580
source IEEE Electronic Library (IEL)
subjects Angles (geometry)
Collaboration
collaborative deep networks
Digital media
Dirichlet problem
Filtration
Geometry
Landmark photo retrieval
Landmarks
Level (quantity)
multi-query expansions
Photography
Poles and towers
Ranking
Retrieval
Robustness
Shape
Social network services
Social networks
Visualization
title Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T17%3A01%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Effective%20Multi-Query%20Expansions:%20Collaborative%20Deep%20Networks%20for%20Robust%20Landmark%20Retrieval&rft.jtitle=IEEE%20transactions%20on%20image%20processing&rft.au=Wang,%20Yang&rft.date=2017-03-01&rft.volume=26&rft.issue=3&rft.spage=1393&rft.epage=1404&rft.pages=1393-1404&rft.issn=1057-7149&rft.eissn=1941-0042&rft.coden=IIPRE4&rft_id=info:doi/10.1109/TIP.2017.2655449&rft_dat=%3Cproquest_RIE%3E1861582580%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1869420062&rft_id=info:pmid/28103558&rft_ieee_id=7823000&rfr_iscdi=true