Content and Context Features for Scene Image Representation
Existing research in scene image classification has focused on either content features (e.g., visual information) or context features (e.g., annotations). As they capture different information about images which can be complementary and useful to discriminate images of different classes, we suppose...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2021-04 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Sitaula, Chiranjibi Aryal, Sunil Xiang, Yong Basnet, Anish Lu, Xuequan |
description | Existing research in scene image classification has focused on either content features (e.g., visual information) or context features (e.g., annotations). As they capture different information about images which can be complementary and useful to discriminate images of different classes, we suppose the fusion of them will improve classification results. In this paper, we propose new techniques to compute content features and context features, and then fuse them together. For content features, we design multi-scale deep features based on background and foreground information in images. For context features, we use annotations of similar images available in the web to design a filter words (codebook). Our experiments in three widely used benchmark scene datasets using support vector machine classifier reveal that our proposed context and content features produce better results than existing context and content features, respectively. The fusion of the proposed two types of features significantly outperform numerous state-of-the-art features. |
doi_str_mv | 10.48550/arxiv.2006.03217 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2006_03217</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2410534993</sourcerecordid><originalsourceid>FETCH-LOGICAL-a523-44ea860693dfc197a5162a0d76715b0aa2414b97ce1ec972e0e514b2229a92ec3</originalsourceid><addsrcrecordid>eNotj0FLw0AQhRdBsNT-AE8ueE6cnd3NZvEkwWqhIGjvYZpMJMUmcZNI_ffG1NM8Hh-P-YS4URCb1Fq4p3Cqv2MESGLQqNyFWKDWKkoN4pVY9f0BADBxaK1eiIesbQZuBklNKed8GuSaaRgD97Jqg3wvuGG5OdIHyzfupnrCaajb5lpcVvTZ8-r_LsVu_bTLXqLt6_Mme9xGZFFHxjClCSRel1WhvCOrEiQoXeKU3QMRGmX23hWsuPAOGdhOBSJ68siFXorb8-xslnehPlL4yf8M89lwIu7ORBfar5H7IT-0Y2imn_JpHKw23mv9C4gwUho</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2410534993</pqid></control><display><type>article</type><title>Content and Context Features for Scene Image Representation</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Sitaula, Chiranjibi ; Aryal, Sunil ; Xiang, Yong ; Basnet, Anish ; Lu, Xuequan</creator><creatorcontrib>Sitaula, Chiranjibi ; Aryal, Sunil ; Xiang, Yong ; Basnet, Anish ; Lu, Xuequan</creatorcontrib><description>Existing research in scene image classification has focused on either content features (e.g., visual information) or context features (e.g., annotations). As they capture different information about images which can be complementary and useful to discriminate images of different classes, we suppose the fusion of them will improve classification results. In this paper, we propose new techniques to compute content features and context features, and then fuse them together. For content features, we design multi-scale deep features based on background and foreground information in images. For context features, we use annotations of similar images available in the web to design a filter words (codebook). Our experiments in three widely used benchmark scene datasets using support vector machine classifier reveal that our proposed context and content features produce better results than existing context and content features, respectively. The fusion of the proposed two types of features significantly outperform numerous state-of-the-art features.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2006.03217</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Annotations ; Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning ; Context ; Image classification ; Support vector machines</subject><ispartof>arXiv.org, 2021-04</ispartof><rights>2021. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,784,885,27925</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2006.03217$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1016/j.knosys.2021.107470$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Sitaula, Chiranjibi</creatorcontrib><creatorcontrib>Aryal, Sunil</creatorcontrib><creatorcontrib>Xiang, Yong</creatorcontrib><creatorcontrib>Basnet, Anish</creatorcontrib><creatorcontrib>Lu, Xuequan</creatorcontrib><title>Content and Context Features for Scene Image Representation</title><title>arXiv.org</title><description>Existing research in scene image classification has focused on either content features (e.g., visual information) or context features (e.g., annotations). As they capture different information about images which can be complementary and useful to discriminate images of different classes, we suppose the fusion of them will improve classification results. In this paper, we propose new techniques to compute content features and context features, and then fuse them together. For content features, we design multi-scale deep features based on background and foreground information in images. For context features, we use annotations of similar images available in the web to design a filter words (codebook). Our experiments in three widely used benchmark scene datasets using support vector machine classifier reveal that our proposed context and content features produce better results than existing context and content features, respectively. The fusion of the proposed two types of features significantly outperform numerous state-of-the-art features.</description><subject>Annotations</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><subject>Context</subject><subject>Image classification</subject><subject>Support vector machines</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GOX</sourceid><recordid>eNotj0FLw0AQhRdBsNT-AE8ueE6cnd3NZvEkwWqhIGjvYZpMJMUmcZNI_ffG1NM8Hh-P-YS4URCb1Fq4p3Cqv2MESGLQqNyFWKDWKkoN4pVY9f0BADBxaK1eiIesbQZuBklNKed8GuSaaRgD97Jqg3wvuGG5OdIHyzfupnrCaajb5lpcVvTZ8-r_LsVu_bTLXqLt6_Mme9xGZFFHxjClCSRel1WhvCOrEiQoXeKU3QMRGmX23hWsuPAOGdhOBSJ68siFXorb8-xslnehPlL4yf8M89lwIu7ORBfar5H7IT-0Y2imn_JpHKw23mv9C4gwUho</recordid><startdate>20210424</startdate><enddate>20210424</enddate><creator>Sitaula, Chiranjibi</creator><creator>Aryal, Sunil</creator><creator>Xiang, Yong</creator><creator>Basnet, Anish</creator><creator>Lu, Xuequan</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210424</creationdate><title>Content and Context Features for Scene Image Representation</title><author>Sitaula, Chiranjibi ; Aryal, Sunil ; Xiang, Yong ; Basnet, Anish ; Lu, Xuequan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a523-44ea860693dfc197a5162a0d76715b0aa2414b97ce1ec972e0e514b2229a92ec3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Annotations</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><topic>Context</topic><topic>Image classification</topic><topic>Support vector machines</topic><toplevel>online_resources</toplevel><creatorcontrib>Sitaula, Chiranjibi</creatorcontrib><creatorcontrib>Aryal, Sunil</creatorcontrib><creatorcontrib>Xiang, Yong</creatorcontrib><creatorcontrib>Basnet, Anish</creatorcontrib><creatorcontrib>Lu, Xuequan</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Sitaula, Chiranjibi</au><au>Aryal, Sunil</au><au>Xiang, Yong</au><au>Basnet, Anish</au><au>Lu, Xuequan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Content and Context Features for Scene Image Representation</atitle><jtitle>arXiv.org</jtitle><date>2021-04-24</date><risdate>2021</risdate><eissn>2331-8422</eissn><abstract>Existing research in scene image classification has focused on either content features (e.g., visual information) or context features (e.g., annotations). As they capture different information about images which can be complementary and useful to discriminate images of different classes, we suppose the fusion of them will improve classification results. In this paper, we propose new techniques to compute content features and context features, and then fuse them together. For content features, we design multi-scale deep features based on background and foreground information in images. For context features, we use annotations of similar images available in the web to design a filter words (codebook). Our experiments in three widely used benchmark scene datasets using support vector machine classifier reveal that our proposed context and content features produce better results than existing context and content features, respectively. The fusion of the proposed two types of features significantly outperform numerous state-of-the-art features.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2006.03217</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2021-04 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_2006_03217 |
source | arXiv.org; Free E- Journals |
subjects | Annotations Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning Context Image classification Support vector machines |
title | Content and Context Features for Scene Image Representation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-25T01%3A53%3A30IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Content%20and%20Context%20Features%20for%20Scene%20Image%20Representation&rft.jtitle=arXiv.org&rft.au=Sitaula,%20Chiranjibi&rft.date=2021-04-24&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2006.03217&rft_dat=%3Cproquest_arxiv%3E2410534993%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2410534993&rft_id=info:pmid/&rfr_iscdi=true |