Learning Multiple Sequence-Based Kernels for Video Concept Detection
Kernel based methods are widely applied to concept and event detection in video. Recently, kernels working on sequences of feature vectors of a video segment have been proposed for this problem, rather than treating feature vectors of individual frames independently. It has been shown that these seq...
Gespeichert in:
1. Verfasser: | |
---|---|
Format: | Tagungsbericht |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 77 |
---|---|
container_issue | |
container_start_page | 73 |
container_title | |
container_volume | |
creator | Bailer, W. |
description | Kernel based methods are widely applied to concept and event detection in video. Recently, kernels working on sequences of feature vectors of a video segment have been proposed for this problem, rather than treating feature vectors of individual frames independently. It has been shown that these sequence-based kernels (based e.g., on the dynamic time warping or edit distance paradigms) outperform methods working on single frames for concepts with inherently dynamic features. Existing work on sequence-based kernels either uses a single type of feature or a fixed combination of the feature vectors of each frame. However, different features (e.g., visual and audio features) may be sampled at different (possibly even irregular) rates, and the optimal alignment between the sequences of features may be different. Multiple kernel learning (MKL) has been applied to similarly structured problems, and we propose MKL for combining different sequence-based kernels on different features for video concept detection. We demonstrate the advantage of the proposed method with experiments on the TRECVID 2011 Semantic Indexing data set. |
doi_str_mv | 10.1109/ISM.2012.22 |
format | Conference Proceeding |
fullrecord | <record><control><sourceid>ieee_6IE</sourceid><recordid>TN_cdi_ieee_primary_6424634</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>6424634</ieee_id><sourcerecordid>6424634</sourcerecordid><originalsourceid>FETCH-LOGICAL-i175t-a75bd31040497f8b895a45524a69cbc7c7d21883b914c4f99aaa29b5b16351233</originalsourceid><addsrcrecordid>eNotjz1PwzAUAI0QErR0YmTxH0h5z5_xCCmFilQMBcRW2c4LMgpJSNKBf08lmG446aRj7AphiQjuZrPbLgWgWApxwmZgjdMqt_r9lM1QGSuVtGDO2WIcPwEAQWpw9oKtSvJDm9oPvj00U-ob4jv6PlAbKbvzI1X8iYaWmpHX3cDfUkUdL7qj7Se-oonilLr2kp3Vvhlp8c85e13fvxSPWfn8sCluyyyh1VPmrQ6VRFCgnK3zkDvtldZCeeNiiDbaSmCey-BQRVU7570XLuiARmoUUs7Z9V83EdG-H9KXH372Rglljn-_yEdI4Q</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Learning Multiple Sequence-Based Kernels for Video Concept Detection</title><source>IEEE Electronic Library (IEL) Conference Proceedings</source><creator>Bailer, W.</creator><creatorcontrib>Bailer, W.</creatorcontrib><description>Kernel based methods are widely applied to concept and event detection in video. Recently, kernels working on sequences of feature vectors of a video segment have been proposed for this problem, rather than treating feature vectors of individual frames independently. It has been shown that these sequence-based kernels (based e.g., on the dynamic time warping or edit distance paradigms) outperform methods working on single frames for concepts with inherently dynamic features. Existing work on sequence-based kernels either uses a single type of feature or a fixed combination of the feature vectors of each frame. However, different features (e.g., visual and audio features) may be sampled at different (possibly even irregular) rates, and the optimal alignment between the sequences of features may be different. Multiple kernel learning (MKL) has been applied to similarly structured problems, and we propose MKL for combining different sequence-based kernels on different features for video concept detection. We demonstrate the advantage of the proposed method with experiments on the TRECVID 2011 Semantic Indexing data set.</description><identifier>ISBN: 1467343706</identifier><identifier>ISBN: 9781467343701</identifier><identifier>EISBN: 076954875X</identifier><identifier>EISBN: 9780769548753</identifier><identifier>DOI: 10.1109/ISM.2012.22</identifier><identifier>CODEN: IEEPAD</identifier><language>eng</language><publisher>IEEE</publisher><subject>feature combination ; Feature extraction ; fusion ; Histograms ; Kernel ; learning ; Multimedia communication ; Streaming media ; Vectors ; Visualization</subject><ispartof>2012 IEEE International Symposium on Multimedia, 2012, p.73-77</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/6424634$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,2058,27925,54920</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/6424634$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Bailer, W.</creatorcontrib><title>Learning Multiple Sequence-Based Kernels for Video Concept Detection</title><title>2012 IEEE International Symposium on Multimedia</title><addtitle>ism</addtitle><description>Kernel based methods are widely applied to concept and event detection in video. Recently, kernels working on sequences of feature vectors of a video segment have been proposed for this problem, rather than treating feature vectors of individual frames independently. It has been shown that these sequence-based kernels (based e.g., on the dynamic time warping or edit distance paradigms) outperform methods working on single frames for concepts with inherently dynamic features. Existing work on sequence-based kernels either uses a single type of feature or a fixed combination of the feature vectors of each frame. However, different features (e.g., visual and audio features) may be sampled at different (possibly even irregular) rates, and the optimal alignment between the sequences of features may be different. Multiple kernel learning (MKL) has been applied to similarly structured problems, and we propose MKL for combining different sequence-based kernels on different features for video concept detection. We demonstrate the advantage of the proposed method with experiments on the TRECVID 2011 Semantic Indexing data set.</description><subject>feature combination</subject><subject>Feature extraction</subject><subject>fusion</subject><subject>Histograms</subject><subject>Kernel</subject><subject>learning</subject><subject>Multimedia communication</subject><subject>Streaming media</subject><subject>Vectors</subject><subject>Visualization</subject><isbn>1467343706</isbn><isbn>9781467343701</isbn><isbn>076954875X</isbn><isbn>9780769548753</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2012</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><sourceid>RIE</sourceid><recordid>eNotjz1PwzAUAI0QErR0YmTxH0h5z5_xCCmFilQMBcRW2c4LMgpJSNKBf08lmG446aRj7AphiQjuZrPbLgWgWApxwmZgjdMqt_r9lM1QGSuVtGDO2WIcPwEAQWpw9oKtSvJDm9oPvj00U-ob4jv6PlAbKbvzI1X8iYaWmpHX3cDfUkUdL7qj7Se-oonilLr2kp3Vvhlp8c85e13fvxSPWfn8sCluyyyh1VPmrQ6VRFCgnK3zkDvtldZCeeNiiDbaSmCey-BQRVU7570XLuiARmoUUs7Z9V83EdG-H9KXH372Rglljn-_yEdI4Q</recordid><startdate>201212</startdate><enddate>201212</enddate><creator>Bailer, W.</creator><general>IEEE</general><scope>6IE</scope><scope>6IL</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIL</scope></search><sort><creationdate>201212</creationdate><title>Learning Multiple Sequence-Based Kernels for Video Concept Detection</title><author>Bailer, W.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i175t-a75bd31040497f8b895a45524a69cbc7c7d21883b914c4f99aaa29b5b16351233</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2012</creationdate><topic>feature combination</topic><topic>Feature extraction</topic><topic>fusion</topic><topic>Histograms</topic><topic>Kernel</topic><topic>learning</topic><topic>Multimedia communication</topic><topic>Streaming media</topic><topic>Vectors</topic><topic>Visualization</topic><toplevel>online_resources</toplevel><creatorcontrib>Bailer, W.</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP All) 1998-Present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Bailer, W.</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Learning Multiple Sequence-Based Kernels for Video Concept Detection</atitle><btitle>2012 IEEE International Symposium on Multimedia</btitle><stitle>ism</stitle><date>2012-12</date><risdate>2012</risdate><spage>73</spage><epage>77</epage><pages>73-77</pages><isbn>1467343706</isbn><isbn>9781467343701</isbn><eisbn>076954875X</eisbn><eisbn>9780769548753</eisbn><coden>IEEPAD</coden><abstract>Kernel based methods are widely applied to concept and event detection in video. Recently, kernels working on sequences of feature vectors of a video segment have been proposed for this problem, rather than treating feature vectors of individual frames independently. It has been shown that these sequence-based kernels (based e.g., on the dynamic time warping or edit distance paradigms) outperform methods working on single frames for concepts with inherently dynamic features. Existing work on sequence-based kernels either uses a single type of feature or a fixed combination of the feature vectors of each frame. However, different features (e.g., visual and audio features) may be sampled at different (possibly even irregular) rates, and the optimal alignment between the sequences of features may be different. Multiple kernel learning (MKL) has been applied to similarly structured problems, and we propose MKL for combining different sequence-based kernels on different features for video concept detection. We demonstrate the advantage of the proposed method with experiments on the TRECVID 2011 Semantic Indexing data set.</abstract><pub>IEEE</pub><doi>10.1109/ISM.2012.22</doi><tpages>5</tpages></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISBN: 1467343706 |
ispartof | 2012 IEEE International Symposium on Multimedia, 2012, p.73-77 |
issn | |
language | eng |
recordid | cdi_ieee_primary_6424634 |
source | IEEE Electronic Library (IEL) Conference Proceedings |
subjects | feature combination Feature extraction fusion Histograms Kernel learning Multimedia communication Streaming media Vectors Visualization |
title | Learning Multiple Sequence-Based Kernels for Video Concept Detection |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T06%3A44%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_6IE&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Learning%20Multiple%20Sequence-Based%20Kernels%20for%20Video%20Concept%20Detection&rft.btitle=2012%20IEEE%20International%20Symposium%20on%20Multimedia&rft.au=Bailer,%20W.&rft.date=2012-12&rft.spage=73&rft.epage=77&rft.pages=73-77&rft.isbn=1467343706&rft.isbn_list=9781467343701&rft.coden=IEEPAD&rft_id=info:doi/10.1109/ISM.2012.22&rft_dat=%3Cieee_6IE%3E6424634%3C/ieee_6IE%3E%3Curl%3E%3C/url%3E&rft.eisbn=076954875X&rft.eisbn_list=9780769548753&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=6424634&rfr_iscdi=true |