Interspeech 2016 - Experiment results for paper "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition"

The files in the dataset correspond to results that have been generated for the submitted Interspeech 2016 paper: "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition" (DOI: 10.21437/Interspeech.2016-480). The paper deals with language model...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Deena, Salil, Hasan, Madina, Mortaza Doulaty Bashkand, Torralba, Oscar Saz, Hain, Thomas
Format: Dataset
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Deena, Salil
Hasan, Madina
Mortaza Doulaty Bashkand
Torralba, Oscar Saz
Hain, Thomas
description The files in the dataset correspond to results that have been generated for the submitted Interspeech 2016 paper: "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition" (DOI: 10.21437/Interspeech.2016-480). The paper deals with language model adaptation for the MGB Challenge 2015 transcription task. The files in the zip file are of three types: - .ctm, which correspond to the output of the automatic speech recognition system and the columns include segment information as well as transcripts of the recognition. - .ctm.filt.sys, which correspond to scoring of the automatic speech recognition system and includes the overall word error rate as well as the number of insertions, deletions and substitutions of the overall system. - .ctm.filt.lur, which provides a more detailed decomposition of the word error rate across multiple genres. The three file types are repeated for all the results described in Table 3 of the paper. The following is a description about the naming convention of the files: rnnlm refers to Recurrent Neural Network Language Model. amrnnlm prefix refers to acoustic model text RNNLM. amlmrnnlm prefix refers to acoustic model + language model text RNNLM. .lattice.rescore suffix refers to results generated with lattice rescoring. .nbest.rescore suffix refers to results generated with nbest rescoring. .baseline refers to baseline RNNLM results. .noadaptation refers to RNNLM results with no adaptation. .genre.finetune refers to genre fine-tuning of the RNNLMs. .genre.adaptationlayer refers to genre adaptation layer fine-tuning of the RNNLMs. .ldafeat.hiddenlayer refers to Latent Dirichlet Allocation (LDA) features at the hidden layer. .genrefeat.hiddenlayer refers to Genre 1-hot auxiliary codes at the hidden layer. .genrefeat.adaptationlayer refers to Genre 1-hot auxiliary codes at the adaptation layer. All three file types are standard outputs that are recognised by the automatic speech recognition community and can be opened using any text editor.
doi_str_mv 10.15131/shef.data.3141910
format Dataset
fullrecord <record><control><sourceid>datacite_PQ8</sourceid><recordid>TN_cdi_datacite_primary_10_15131_shef_data_3141910</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_15131_shef_data_3141910</sourcerecordid><originalsourceid>FETCH-datacite_primary_10_15131_shef_data_31419103</originalsourceid><addsrcrecordid>eNqdj8FuwjAMhnPhMMFeYCeLe7uaAhJHQLBNWjkA98g0LkQqSZQYaXuQve_awRPsZMnW9_v_lHrBIscZlviaLtzkhoTyEqe4wOJJ_Xw44ZgCc32BSYFzyGDzFTjaKzuByOnWSoLGRwjUrWG89teTddadYcskt8hAzkDlDbfZihIbWBoKQmK9A9_Afrf7rO4JVZdlszd2HbSKnkxNSeBwf77n2p-d7bHxSA0aahM_P-ZQTbab4_o968vXVliHriDFb42F_nPTvZvuz_rhVv4L-gWd62F4</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>dataset</recordtype></control><display><type>dataset</type><title>Interspeech 2016 - Experiment results for paper "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition"</title><source>DataCite</source><creator>Deena, Salil ; Hasan, Madina ; Mortaza Doulaty Bashkand ; Torralba, Oscar Saz ; Hain, Thomas</creator><creatorcontrib>Deena, Salil ; Hasan, Madina ; Mortaza Doulaty Bashkand ; Torralba, Oscar Saz ; Hain, Thomas</creatorcontrib><description>The files in the dataset correspond to results that have been generated for the submitted Interspeech 2016 paper: "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition" (DOI: 10.21437/Interspeech.2016-480). The paper deals with language model adaptation for the MGB Challenge 2015 transcription task. The files in the zip file are of three types: - .ctm, which correspond to the output of the automatic speech recognition system and the columns include segment information as well as transcripts of the recognition. - .ctm.filt.sys, which correspond to scoring of the automatic speech recognition system and includes the overall word error rate as well as the number of insertions, deletions and substitutions of the overall system. - .ctm.filt.lur, which provides a more detailed decomposition of the word error rate across multiple genres. The three file types are repeated for all the results described in Table 3 of the paper. The following is a description about the naming convention of the files: rnnlm refers to Recurrent Neural Network Language Model. amrnnlm prefix refers to acoustic model text RNNLM. amlmrnnlm prefix refers to acoustic model + language model text RNNLM. .lattice.rescore suffix refers to results generated with lattice rescoring. .nbest.rescore suffix refers to results generated with nbest rescoring. .baseline refers to baseline RNNLM results. .noadaptation refers to RNNLM results with no adaptation. .genre.finetune refers to genre fine-tuning of the RNNLMs. .genre.adaptationlayer refers to genre adaptation layer fine-tuning of the RNNLMs. .ldafeat.hiddenlayer refers to Latent Dirichlet Allocation (LDA) features at the hidden layer. .genrefeat.hiddenlayer refers to Genre 1-hot auxiliary codes at the hidden layer. .genrefeat.adaptationlayer refers to Genre 1-hot auxiliary codes at the adaptation layer. All three file types are standard outputs that are recognised by the automatic speech recognition community and can be opened using any text editor.</description><identifier>DOI: 10.15131/shef.data.3141910</identifier><language>eng</language><publisher>The University of Sheffield</publisher><subject>FOS: Computer and information sciences ; Natural Language Processing</subject><creationdate>2016</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,1893</link.rule.ids><linktorsrc>$$Uhttps://commons.datacite.org/doi.org/10.15131/shef.data.3141910$$EView_record_in_DataCite.org$$FView_record_in_$$GDataCite.org$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>Deena, Salil</creatorcontrib><creatorcontrib>Hasan, Madina</creatorcontrib><creatorcontrib>Mortaza Doulaty Bashkand</creatorcontrib><creatorcontrib>Torralba, Oscar Saz</creatorcontrib><creatorcontrib>Hain, Thomas</creatorcontrib><title>Interspeech 2016 - Experiment results for paper "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition"</title><description>The files in the dataset correspond to results that have been generated for the submitted Interspeech 2016 paper: "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition" (DOI: 10.21437/Interspeech.2016-480). The paper deals with language model adaptation for the MGB Challenge 2015 transcription task. The files in the zip file are of three types: - .ctm, which correspond to the output of the automatic speech recognition system and the columns include segment information as well as transcripts of the recognition. - .ctm.filt.sys, which correspond to scoring of the automatic speech recognition system and includes the overall word error rate as well as the number of insertions, deletions and substitutions of the overall system. - .ctm.filt.lur, which provides a more detailed decomposition of the word error rate across multiple genres. The three file types are repeated for all the results described in Table 3 of the paper. The following is a description about the naming convention of the files: rnnlm refers to Recurrent Neural Network Language Model. amrnnlm prefix refers to acoustic model text RNNLM. amlmrnnlm prefix refers to acoustic model + language model text RNNLM. .lattice.rescore suffix refers to results generated with lattice rescoring. .nbest.rescore suffix refers to results generated with nbest rescoring. .baseline refers to baseline RNNLM results. .noadaptation refers to RNNLM results with no adaptation. .genre.finetune refers to genre fine-tuning of the RNNLMs. .genre.adaptationlayer refers to genre adaptation layer fine-tuning of the RNNLMs. .ldafeat.hiddenlayer refers to Latent Dirichlet Allocation (LDA) features at the hidden layer. .genrefeat.hiddenlayer refers to Genre 1-hot auxiliary codes at the hidden layer. .genrefeat.adaptationlayer refers to Genre 1-hot auxiliary codes at the adaptation layer. All three file types are standard outputs that are recognised by the automatic speech recognition community and can be opened using any text editor.</description><subject>FOS: Computer and information sciences</subject><subject>Natural Language Processing</subject><fulltext>true</fulltext><rsrctype>dataset</rsrctype><creationdate>2016</creationdate><recordtype>dataset</recordtype><sourceid>PQ8</sourceid><recordid>eNqdj8FuwjAMhnPhMMFeYCeLe7uaAhJHQLBNWjkA98g0LkQqSZQYaXuQve_awRPsZMnW9_v_lHrBIscZlviaLtzkhoTyEqe4wOJJ_Xw44ZgCc32BSYFzyGDzFTjaKzuByOnWSoLGRwjUrWG89teTddadYcskt8hAzkDlDbfZihIbWBoKQmK9A9_Afrf7rO4JVZdlszd2HbSKnkxNSeBwf77n2p-d7bHxSA0aahM_P-ZQTbab4_o968vXVliHriDFb42F_nPTvZvuz_rhVv4L-gWd62F4</recordid><startdate>20160401</startdate><enddate>20160401</enddate><creator>Deena, Salil</creator><creator>Hasan, Madina</creator><creator>Mortaza Doulaty Bashkand</creator><creator>Torralba, Oscar Saz</creator><creator>Hain, Thomas</creator><general>The University of Sheffield</general><scope>DYCCY</scope><scope>PQ8</scope></search><sort><creationdate>20160401</creationdate><title>Interspeech 2016 - Experiment results for paper "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition"</title><author>Deena, Salil ; Hasan, Madina ; Mortaza Doulaty Bashkand ; Torralba, Oscar Saz ; Hain, Thomas</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-datacite_primary_10_15131_shef_data_31419103</frbrgroupid><rsrctype>datasets</rsrctype><prefilter>datasets</prefilter><language>eng</language><creationdate>2016</creationdate><topic>FOS: Computer and information sciences</topic><topic>Natural Language Processing</topic><toplevel>online_resources</toplevel><creatorcontrib>Deena, Salil</creatorcontrib><creatorcontrib>Hasan, Madina</creatorcontrib><creatorcontrib>Mortaza Doulaty Bashkand</creatorcontrib><creatorcontrib>Torralba, Oscar Saz</creatorcontrib><creatorcontrib>Hain, Thomas</creatorcontrib><collection>DataCite (Open Access)</collection><collection>DataCite</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Deena, Salil</au><au>Hasan, Madina</au><au>Mortaza Doulaty Bashkand</au><au>Torralba, Oscar Saz</au><au>Hain, Thomas</au><format>book</format><genre>unknown</genre><ristype>DATA</ristype><title>Interspeech 2016 - Experiment results for paper "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition"</title><date>2016-04-01</date><risdate>2016</risdate><abstract>The files in the dataset correspond to results that have been generated for the submitted Interspeech 2016 paper: "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition" (DOI: 10.21437/Interspeech.2016-480). The paper deals with language model adaptation for the MGB Challenge 2015 transcription task. The files in the zip file are of three types: - .ctm, which correspond to the output of the automatic speech recognition system and the columns include segment information as well as transcripts of the recognition. - .ctm.filt.sys, which correspond to scoring of the automatic speech recognition system and includes the overall word error rate as well as the number of insertions, deletions and substitutions of the overall system. - .ctm.filt.lur, which provides a more detailed decomposition of the word error rate across multiple genres. The three file types are repeated for all the results described in Table 3 of the paper. The following is a description about the naming convention of the files: rnnlm refers to Recurrent Neural Network Language Model. amrnnlm prefix refers to acoustic model text RNNLM. amlmrnnlm prefix refers to acoustic model + language model text RNNLM. .lattice.rescore suffix refers to results generated with lattice rescoring. .nbest.rescore suffix refers to results generated with nbest rescoring. .baseline refers to baseline RNNLM results. .noadaptation refers to RNNLM results with no adaptation. .genre.finetune refers to genre fine-tuning of the RNNLMs. .genre.adaptationlayer refers to genre adaptation layer fine-tuning of the RNNLMs. .ldafeat.hiddenlayer refers to Latent Dirichlet Allocation (LDA) features at the hidden layer. .genrefeat.hiddenlayer refers to Genre 1-hot auxiliary codes at the hidden layer. .genrefeat.adaptationlayer refers to Genre 1-hot auxiliary codes at the adaptation layer. All three file types are standard outputs that are recognised by the automatic speech recognition community and can be opened using any text editor.</abstract><pub>The University of Sheffield</pub><doi>10.15131/shef.data.3141910</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.15131/shef.data.3141910
ispartof
issn
language eng
recordid cdi_datacite_primary_10_15131_shef_data_3141910
source DataCite
subjects FOS: Computer and information sciences
Natural Language Processing
title Interspeech 2016 - Experiment results for paper "Combining Feature and Model-Based Adaptation of RNNLMs for Multi-Genre Broadcast Speech Recognition"
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T08%3A32%3A15IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-datacite_PQ8&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=unknown&rft.au=Deena,%20Salil&rft.date=2016-04-01&rft_id=info:doi/10.15131/shef.data.3141910&rft_dat=%3Cdatacite_PQ8%3E10_15131_shef_data_3141910%3C/datacite_PQ8%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true