Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval
Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on image processing 2017-03, Vol.26 (3), p.1393-1404 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1404 |
---|---|
container_issue | 3 |
container_start_page | 1393 |
container_title | IEEE transactions on image processing |
container_volume | 26 |
creator | Wang, Yang Lin, Xuemin Wu, Lin Zhang, Wenjie |
description | Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1]. |
doi_str_mv | 10.1109/TIP.2017.2655449 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_1861582580</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>7823000</ieee_id><sourcerecordid>1861582580</sourcerecordid><originalsourceid>FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</originalsourceid><addsrcrecordid>eNpdkMtP40AMh0erRQsU7iuthCJx4ZJizysTbqgUqFSeYk97iPJwpECaKTMJj_-eKS0c9mRL_mz9_DH2G2GMCOnxw-x2zAGTMddKSZn-YDuYSowBJP8ZelBJnKBMt9mu948AKBXqX2ybGwShlNlh_6Z1TWXfvFB0NbR9E98N5N6j6dsy73xjO38STWzb5oV1-Sd1RrSMrql_te7JR7V10b0tBt9H87yrFrl7iu6pdw295O0e26rz1tP-po7Y3_Ppw-Qynt9czCan87gUMunjRJTATa4JtUhVWSleFigKXXMsjdKJFkaqSinBiScK0xCdijrl3JjaqMqIETta3106-zyQ77NF40sKqTuyg8_QaFSGKwMBPfwPfbSD60K6FZVKDqB5oGBNlc5676jOlq4Jv71nCNlKfBbEZyvx2UZ8WDnYHB6KBVXfC1-mA_BnDTRE9D1ODBcAID4Ag6OFJw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1869420062</pqid></control><display><type>article</type><title>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</title><source>IEEE Electronic Library (IEL)</source><creator>Wang, Yang ; Lin, Xuemin ; Wu, Lin ; Zhang, Wenjie</creator><creatorcontrib>Wang, Yang ; Lin, Xuemin ; Wu, Lin ; Zhang, Wenjie</creatorcontrib><description>Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].</description><identifier>ISSN: 1057-7149</identifier><identifier>EISSN: 1941-0042</identifier><identifier>DOI: 10.1109/TIP.2017.2655449</identifier><identifier>PMID: 28103558</identifier><identifier>CODEN: IIPRE4</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Angles (geometry) ; Collaboration ; collaborative deep networks ; Digital media ; Dirichlet problem ; Filtration ; Geometry ; Landmark photo retrieval ; Landmarks ; Level (quantity) ; multi-query expansions ; Photography ; Poles and towers ; Ranking ; Retrieval ; Robustness ; Shape ; Social network services ; Social networks ; Visualization</subject><ispartof>IEEE transactions on image processing, 2017-03, Vol.26 (3), p.1393-1404</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2017</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</citedby><cites>FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</cites><orcidid>0000-0003-1029-9280</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/7823000$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27923,27924,54757</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/7823000$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/28103558$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Wang, Yang</creatorcontrib><creatorcontrib>Lin, Xuemin</creatorcontrib><creatorcontrib>Wu, Lin</creatorcontrib><creatorcontrib>Zhang, Wenjie</creatorcontrib><title>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</title><title>IEEE transactions on image processing</title><addtitle>TIP</addtitle><addtitle>IEEE Trans Image Process</addtitle><description>Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].</description><subject>Angles (geometry)</subject><subject>Collaboration</subject><subject>collaborative deep networks</subject><subject>Digital media</subject><subject>Dirichlet problem</subject><subject>Filtration</subject><subject>Geometry</subject><subject>Landmark photo retrieval</subject><subject>Landmarks</subject><subject>Level (quantity)</subject><subject>multi-query expansions</subject><subject>Photography</subject><subject>Poles and towers</subject><subject>Ranking</subject><subject>Retrieval</subject><subject>Robustness</subject><subject>Shape</subject><subject>Social network services</subject><subject>Social networks</subject><subject>Visualization</subject><issn>1057-7149</issn><issn>1941-0042</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2017</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpdkMtP40AMh0erRQsU7iuthCJx4ZJizysTbqgUqFSeYk97iPJwpECaKTMJj_-eKS0c9mRL_mz9_DH2G2GMCOnxw-x2zAGTMddKSZn-YDuYSowBJP8ZelBJnKBMt9mu948AKBXqX2ybGwShlNlh_6Z1TWXfvFB0NbR9E98N5N6j6dsy73xjO38STWzb5oV1-Sd1RrSMrql_te7JR7V10b0tBt9H87yrFrl7iu6pdw295O0e26rz1tP-po7Y3_Ppw-Qynt9czCan87gUMunjRJTATa4JtUhVWSleFigKXXMsjdKJFkaqSinBiScK0xCdijrl3JjaqMqIETta3106-zyQ77NF40sKqTuyg8_QaFSGKwMBPfwPfbSD60K6FZVKDqB5oGBNlc5676jOlq4Jv71nCNlKfBbEZyvx2UZ8WDnYHB6KBVXfC1-mA_BnDTRE9D1ODBcAID4Ag6OFJw</recordid><startdate>20170301</startdate><enddate>20170301</enddate><creator>Wang, Yang</creator><creator>Lin, Xuemin</creator><creator>Wu, Lin</creator><creator>Zhang, Wenjie</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-1029-9280</orcidid></search><sort><creationdate>20170301</creationdate><title>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</title><author>Wang, Yang ; Lin, Xuemin ; Wu, Lin ; Zhang, Wenjie</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c347t-73c028a6e16395cd52cb13b6f21c856763845d5532e27519103ebf92288f85d83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2017</creationdate><topic>Angles (geometry)</topic><topic>Collaboration</topic><topic>collaborative deep networks</topic><topic>Digital media</topic><topic>Dirichlet problem</topic><topic>Filtration</topic><topic>Geometry</topic><topic>Landmark photo retrieval</topic><topic>Landmarks</topic><topic>Level (quantity)</topic><topic>multi-query expansions</topic><topic>Photography</topic><topic>Poles and towers</topic><topic>Ranking</topic><topic>Retrieval</topic><topic>Robustness</topic><topic>Shape</topic><topic>Social network services</topic><topic>Social networks</topic><topic>Visualization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Yang</creatorcontrib><creatorcontrib>Lin, Xuemin</creatorcontrib><creatorcontrib>Wu, Lin</creatorcontrib><creatorcontrib>Zhang, Wenjie</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transactions on image processing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wang, Yang</au><au>Lin, Xuemin</au><au>Wu, Lin</au><au>Zhang, Wenjie</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval</atitle><jtitle>IEEE transactions on image processing</jtitle><stitle>TIP</stitle><addtitle>IEEE Trans Image Process</addtitle><date>2017-03-01</date><risdate>2017</risdate><volume>26</volume><issue>3</issue><spage>1393</spage><epage>1404</epage><pages>1393-1404</pages><issn>1057-7149</issn><eissn>1941-0042</eissn><coden>IIPRE4</coden><abstract>Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top-k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].</abstract><cop>United States</cop><pub>IEEE</pub><pmid>28103558</pmid><doi>10.1109/TIP.2017.2655449</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0003-1029-9280</orcidid></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 1057-7149 |
ispartof | IEEE transactions on image processing, 2017-03, Vol.26 (3), p.1393-1404 |
issn | 1057-7149 1941-0042 |
language | eng |
recordid | cdi_proquest_miscellaneous_1861582580 |
source | IEEE Electronic Library (IEL) |
subjects | Angles (geometry) Collaboration collaborative deep networks Digital media Dirichlet problem Filtration Geometry Landmark photo retrieval Landmarks Level (quantity) multi-query expansions Photography Poles and towers Ranking Retrieval Robustness Shape Social network services Social networks Visualization |
title | Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T17%3A01%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Effective%20Multi-Query%20Expansions:%20Collaborative%20Deep%20Networks%20for%20Robust%20Landmark%20Retrieval&rft.jtitle=IEEE%20transactions%20on%20image%20processing&rft.au=Wang,%20Yang&rft.date=2017-03-01&rft.volume=26&rft.issue=3&rft.spage=1393&rft.epage=1404&rft.pages=1393-1404&rft.issn=1057-7149&rft.eissn=1941-0042&rft.coden=IIPRE4&rft_id=info:doi/10.1109/TIP.2017.2655449&rft_dat=%3Cproquest_RIE%3E1861582580%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1869420062&rft_id=info:pmid/28103558&rft_ieee_id=7823000&rfr_iscdi=true |