Fischlar-TRECVid-2004: combined text- and image-based searching of video archives

The Fischlar-TRECVid-2004 system was developed for Dublin City University's participation in the 2004 TRECVid video information retrieval benchmarking activity. The system allows search and retrieval of video shots from over 60 hours of content. The shot retrieval engine employed is based on a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: O'Connor, N.A., Hyowon Lee, Smeaton, A.F., Jones, G.J.F., Cooke, E., Le Borgne, H., Gurrin, C.
Format: Tagungsbericht
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 4 pp.
container_title
container_volume
creator O'Connor, N.A.
Hyowon Lee
Smeaton, A.F.
Jones, G.J.F.
Cooke, E.
Le Borgne, H.
Gurrin, C.
description The Fischlar-TRECVid-2004 system was developed for Dublin City University's participation in the 2004 TRECVid video information retrieval benchmarking activity. The system allows search and retrieval of video shots from over 60 hours of content. The shot retrieval engine employed is based on a combination of query text matched against spoken dialogue combined with image-image matching where a still image (sourced externally), or a keyframe (from within the video archive itself), is matched against all keyframes in the video archive. Three separate text retrieval engines are employed for closed caption text, automatic speech recognition and video OCR. Visual shot matching is primarily based on MPEG-7 low-level descriptors. The system supports relevance feedback at the shot level enabling augmentation and refinement using relevant shots located by the user. Two variants of the system were developed, one that supports both text- and image-based searching and one that supports image only search. A user evaluation experiment compared the use of the two systems. Results show that while the system combining text- and image-based searching achieves greater retrieval effectiveness, users make more varied and extensive queries with the image only based searching version
doi_str_mv 10.1109/ISCAS.2006.1693029
format Conference Proceeding
fullrecord <record><control><sourceid>ieee_6IE</sourceid><recordid>TN_cdi_ieee_primary_1693029</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>1693029</ieee_id><sourcerecordid>1693029</sourcerecordid><originalsourceid>FETCH-LOGICAL-i649-f3ab72475a12b41bc919227d66ddfe2a5b346f68f19f30dd34b945395426f1503</originalsourceid><addsrcrecordid>eNotUM1KAzEYDP6Aa-0L6CUvkPXL7268laXVQkG0i9eSbJI20u7KphR9e4N2LsMMwzAMQvcUSkpBPy7XzWxdMgBVUqU5MH2BCkZlTahk8hLdQlUD17zW-goVwCpKRE7doGlKn5AhZNZQoLdFTN1ub0bSvs-bj-hILhVPuBsONvbe4aP_PhJseofjwWw9sSZlN3kzdrvYb_EQ8Ck6P-A_4-TTHboOZp_89MwT1C7mbfNCVq_Py2a2IlEJTQI3tmKikoYyK6jtNNWMVU4p54JnRlouVFB1oDpwcI4Lq4XkWgqmApXAJ-jhvzZ67zdfY143_mzOZ_BfXXNO8Q</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>conference_proceeding</recordtype></control><display><type>conference_proceeding</type><title>Fischlar-TRECVid-2004: combined text- and image-based searching of video archives</title><source>IEEE Electronic Library (IEL) Conference Proceedings</source><creator>O'Connor, N.A. ; Hyowon Lee ; Smeaton, A.F. ; Jones, G.J.F. ; Cooke, E. ; Le Borgne, H. ; Gurrin, C.</creator><creatorcontrib>O'Connor, N.A. ; Hyowon Lee ; Smeaton, A.F. ; Jones, G.J.F. ; Cooke, E. ; Le Borgne, H. ; Gurrin, C.</creatorcontrib><description>The Fischlar-TRECVid-2004 system was developed for Dublin City University's participation in the 2004 TRECVid video information retrieval benchmarking activity. The system allows search and retrieval of video shots from over 60 hours of content. The shot retrieval engine employed is based on a combination of query text matched against spoken dialogue combined with image-image matching where a still image (sourced externally), or a keyframe (from within the video archive itself), is matched against all keyframes in the video archive. Three separate text retrieval engines are employed for closed caption text, automatic speech recognition and video OCR. Visual shot matching is primarily based on MPEG-7 low-level descriptors. The system supports relevance feedback at the shot level enabling augmentation and refinement using relevant shots located by the user. Two variants of the system were developed, one that supports both text- and image-based searching and one that supports image only search. A user evaluation experiment compared the use of the two systems. Results show that while the system combining text- and image-based searching achieves greater retrieval effectiveness, users make more varied and extensive queries with the image only based searching version</description><identifier>ISSN: 0271-4302</identifier><identifier>ISBN: 0780393899</identifier><identifier>ISBN: 9780780393899</identifier><identifier>EISSN: 2158-1525</identifier><identifier>DOI: 10.1109/ISCAS.2006.1693029</identifier><language>eng</language><publisher>IEEE</publisher><subject>Automatic speech recognition ; Broadcasting ; Content based retrieval ; Engines ; Feedback ; Information retrieval ; MPEG 7 Standard ; Optical character recognition software ; System testing ; Time factors</subject><ispartof>2006 IEEE International Symposium on Circuits and Systems (ISCAS), 2006, p.4 pp.</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/1693029$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>309,310,780,784,789,790,2058,4050,4051,27925,54920</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/1693029$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>O'Connor, N.A.</creatorcontrib><creatorcontrib>Hyowon Lee</creatorcontrib><creatorcontrib>Smeaton, A.F.</creatorcontrib><creatorcontrib>Jones, G.J.F.</creatorcontrib><creatorcontrib>Cooke, E.</creatorcontrib><creatorcontrib>Le Borgne, H.</creatorcontrib><creatorcontrib>Gurrin, C.</creatorcontrib><title>Fischlar-TRECVid-2004: combined text- and image-based searching of video archives</title><title>2006 IEEE International Symposium on Circuits and Systems (ISCAS)</title><addtitle>ISCAS</addtitle><description>The Fischlar-TRECVid-2004 system was developed for Dublin City University's participation in the 2004 TRECVid video information retrieval benchmarking activity. The system allows search and retrieval of video shots from over 60 hours of content. The shot retrieval engine employed is based on a combination of query text matched against spoken dialogue combined with image-image matching where a still image (sourced externally), or a keyframe (from within the video archive itself), is matched against all keyframes in the video archive. Three separate text retrieval engines are employed for closed caption text, automatic speech recognition and video OCR. Visual shot matching is primarily based on MPEG-7 low-level descriptors. The system supports relevance feedback at the shot level enabling augmentation and refinement using relevant shots located by the user. Two variants of the system were developed, one that supports both text- and image-based searching and one that supports image only search. A user evaluation experiment compared the use of the two systems. Results show that while the system combining text- and image-based searching achieves greater retrieval effectiveness, users make more varied and extensive queries with the image only based searching version</description><subject>Automatic speech recognition</subject><subject>Broadcasting</subject><subject>Content based retrieval</subject><subject>Engines</subject><subject>Feedback</subject><subject>Information retrieval</subject><subject>MPEG 7 Standard</subject><subject>Optical character recognition software</subject><subject>System testing</subject><subject>Time factors</subject><issn>0271-4302</issn><issn>2158-1525</issn><isbn>0780393899</isbn><isbn>9780780393899</isbn><fulltext>true</fulltext><rsrctype>conference_proceeding</rsrctype><creationdate>2006</creationdate><recordtype>conference_proceeding</recordtype><sourceid>6IE</sourceid><sourceid>RIE</sourceid><recordid>eNotUM1KAzEYDP6Aa-0L6CUvkPXL7268laXVQkG0i9eSbJI20u7KphR9e4N2LsMMwzAMQvcUSkpBPy7XzWxdMgBVUqU5MH2BCkZlTahk8hLdQlUD17zW-goVwCpKRE7doGlKn5AhZNZQoLdFTN1ub0bSvs-bj-hILhVPuBsONvbe4aP_PhJseofjwWw9sSZlN3kzdrvYb_EQ8Ck6P-A_4-TTHboOZp_89MwT1C7mbfNCVq_Py2a2IlEJTQI3tmKikoYyK6jtNNWMVU4p54JnRlouVFB1oDpwcI4Lq4XkWgqmApXAJ-jhvzZ67zdfY143_mzOZ_BfXXNO8Q</recordid><startdate>2006</startdate><enddate>2006</enddate><creator>O'Connor, N.A.</creator><creator>Hyowon Lee</creator><creator>Smeaton, A.F.</creator><creator>Jones, G.J.F.</creator><creator>Cooke, E.</creator><creator>Le Borgne, H.</creator><creator>Gurrin, C.</creator><general>IEEE</general><scope>6IE</scope><scope>6IH</scope><scope>CBEJK</scope><scope>RIE</scope><scope>RIO</scope></search><sort><creationdate>2006</creationdate><title>Fischlar-TRECVid-2004: combined text- and image-based searching of video archives</title><author>O'Connor, N.A. ; Hyowon Lee ; Smeaton, A.F. ; Jones, G.J.F. ; Cooke, E. ; Le Borgne, H. ; Gurrin, C.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-i649-f3ab72475a12b41bc919227d66ddfe2a5b346f68f19f30dd34b945395426f1503</frbrgroupid><rsrctype>conference_proceedings</rsrctype><prefilter>conference_proceedings</prefilter><language>eng</language><creationdate>2006</creationdate><topic>Automatic speech recognition</topic><topic>Broadcasting</topic><topic>Content based retrieval</topic><topic>Engines</topic><topic>Feedback</topic><topic>Information retrieval</topic><topic>MPEG 7 Standard</topic><topic>Optical character recognition software</topic><topic>System testing</topic><topic>Time factors</topic><toplevel>online_resources</toplevel><creatorcontrib>O'Connor, N.A.</creatorcontrib><creatorcontrib>Hyowon Lee</creatorcontrib><creatorcontrib>Smeaton, A.F.</creatorcontrib><creatorcontrib>Jones, G.J.F.</creatorcontrib><creatorcontrib>Cooke, E.</creatorcontrib><creatorcontrib>Le Borgne, H.</creatorcontrib><creatorcontrib>Gurrin, C.</creatorcontrib><collection>IEEE Electronic Library (IEL) Conference Proceedings</collection><collection>IEEE Proceedings Order Plan (POP) 1998-present by volume</collection><collection>IEEE Xplore All Conference Proceedings</collection><collection>IEEE Electronic Library (IEL)</collection><collection>IEEE Proceedings Order Plans (POP) 1998-present</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>O'Connor, N.A.</au><au>Hyowon Lee</au><au>Smeaton, A.F.</au><au>Jones, G.J.F.</au><au>Cooke, E.</au><au>Le Borgne, H.</au><au>Gurrin, C.</au><format>book</format><genre>proceeding</genre><ristype>CONF</ristype><atitle>Fischlar-TRECVid-2004: combined text- and image-based searching of video archives</atitle><btitle>2006 IEEE International Symposium on Circuits and Systems (ISCAS)</btitle><stitle>ISCAS</stitle><date>2006</date><risdate>2006</risdate><spage>4 pp.</spage><pages>4 pp.-</pages><issn>0271-4302</issn><eissn>2158-1525</eissn><isbn>0780393899</isbn><isbn>9780780393899</isbn><abstract>The Fischlar-TRECVid-2004 system was developed for Dublin City University's participation in the 2004 TRECVid video information retrieval benchmarking activity. The system allows search and retrieval of video shots from over 60 hours of content. The shot retrieval engine employed is based on a combination of query text matched against spoken dialogue combined with image-image matching where a still image (sourced externally), or a keyframe (from within the video archive itself), is matched against all keyframes in the video archive. Three separate text retrieval engines are employed for closed caption text, automatic speech recognition and video OCR. Visual shot matching is primarily based on MPEG-7 low-level descriptors. The system supports relevance feedback at the shot level enabling augmentation and refinement using relevant shots located by the user. Two variants of the system were developed, one that supports both text- and image-based searching and one that supports image only search. A user evaluation experiment compared the use of the two systems. Results show that while the system combining text- and image-based searching achieves greater retrieval effectiveness, users make more varied and extensive queries with the image only based searching version</abstract><pub>IEEE</pub><doi>10.1109/ISCAS.2006.1693029</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 0271-4302
ispartof 2006 IEEE International Symposium on Circuits and Systems (ISCAS), 2006, p.4 pp.
issn 0271-4302
2158-1525
language eng
recordid cdi_ieee_primary_1693029
source IEEE Electronic Library (IEL) Conference Proceedings
subjects Automatic speech recognition
Broadcasting
Content based retrieval
Engines
Feedback
Information retrieval
MPEG 7 Standard
Optical character recognition software
System testing
Time factors
title Fischlar-TRECVid-2004: combined text- and image-based searching of video archives
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-22T06%3A13%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_6IE&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=proceeding&rft.atitle=Fischlar-TRECVid-2004:%20combined%20text-%20and%20image-based%20searching%20of%20video%20archives&rft.btitle=2006%20IEEE%20International%20Symposium%20on%20Circuits%20and%20Systems%20(ISCAS)&rft.au=O'Connor,%20N.A.&rft.date=2006&rft.spage=4%20pp.&rft.pages=4%20pp.-&rft.issn=0271-4302&rft.eissn=2158-1525&rft.isbn=0780393899&rft.isbn_list=9780780393899&rft_id=info:doi/10.1109/ISCAS.2006.1693029&rft_dat=%3Cieee_6IE%3E1693029%3C/ieee_6IE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=1693029&rfr_iscdi=true