Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures

Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audio...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of the Acoustical Society of America 2021-10, Vol.150 (4), p.A305-A305
Hauptverfasser: Venezia, Jonathan H., Whittle, Nicole, Herrera Ortiz, Christian, Leek, Marjorie R., Barcenas, Caleb, Lee, Grace
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page A305
container_issue 4
container_start_page A305
container_title The Journal of the Acoustical Society of America
container_volume 150
creator Venezia, Jonathan H.
Whittle, Nicole
Herrera Ortiz, Christian
Leek, Marjorie R.
Barcenas, Caleb
Lee, Grace
description Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.
doi_str_mv 10.1121/10.0008375
format Article
fullrecord <record><control><sourceid>scitation_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1121_10_0008375</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>jasa</sourcerecordid><originalsourceid>FETCH-LOGICAL-c715-570190dbfe6e3e1401c1046a1b87b6c28256635d02201fd51eb85493f618854d3</originalsourceid><addsrcrecordid>eNp9kD9PwzAUxC0EEqWw8Ak8gwJ-TuwkI6qgIFWCoXvl2M-toYkjO0HqwHfH_cPKdPdOv3vDEXIL7AGAw2NSxliVl-KMTEBwllWCF-dkklLIilrKS3IV42c6RZXXE_LzEdA4PTjfUW9p7BH1hgbUft25Q-o62ij9tQ5-7AztvItIVXLatz0Orlv_lWzwLY1jH9SwCRg3fmuoGo0bfNidGoef30hbVHFMzDW5sGob8eakU7J8eV7OXrPF-_xt9rTIdAkiEyWDmpnGosQcoWCggRVSQVOVjdS84kLKXBjGOQNrBGBTiaLOrYQqGZNPyd3xrQ4-xoB21QfXqrBbAVvtd9vrabcE3x_hqN2g9hP8R_8CuTtvQw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</title><source>AIP Journals Complete</source><source>Alma/SFX Local Collection</source><source>AIP Acoustical Society of America</source><creator>Venezia, Jonathan H. ; Whittle, Nicole ; Herrera Ortiz, Christian ; Leek, Marjorie R. ; Barcenas, Caleb ; Lee, Grace</creator><creatorcontrib>Venezia, Jonathan H. ; Whittle, Nicole ; Herrera Ortiz, Christian ; Leek, Marjorie R. ; Barcenas, Caleb ; Lee, Grace</creatorcontrib><description>Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.</description><identifier>ISSN: 0001-4966</identifier><identifier>EISSN: 1520-8524</identifier><identifier>DOI: 10.1121/10.0008375</identifier><identifier>CODEN: JASMAN</identifier><language>eng</language><ispartof>The Journal of the Acoustical Society of America, 2021-10, Vol.150 (4), p.A305-A305</ispartof><rights>Acoustical Society of America</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://pubs.aip.org/jasa/article-lookup/doi/10.1121/10.0008375$$EHTML$$P50$$Gscitation$$H</linktohtml><link.rule.ids>207,208,314,776,780,790,1559,4498,27901,27902,76126</link.rule.ids></links><search><creatorcontrib>Venezia, Jonathan H.</creatorcontrib><creatorcontrib>Whittle, Nicole</creatorcontrib><creatorcontrib>Herrera Ortiz, Christian</creatorcontrib><creatorcontrib>Leek, Marjorie R.</creatorcontrib><creatorcontrib>Barcenas, Caleb</creatorcontrib><creatorcontrib>Lee, Grace</creatorcontrib><title>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</title><title>The Journal of the Acoustical Society of America</title><description>Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.</description><issn>0001-4966</issn><issn>1520-8524</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kD9PwzAUxC0EEqWw8Ak8gwJ-TuwkI6qgIFWCoXvl2M-toYkjO0HqwHfH_cPKdPdOv3vDEXIL7AGAw2NSxliVl-KMTEBwllWCF-dkklLIilrKS3IV42c6RZXXE_LzEdA4PTjfUW9p7BH1hgbUft25Q-o62ij9tQ5-7AztvItIVXLatz0Orlv_lWzwLY1jH9SwCRg3fmuoGo0bfNidGoef30hbVHFMzDW5sGob8eakU7J8eV7OXrPF-_xt9rTIdAkiEyWDmpnGosQcoWCggRVSQVOVjdS84kLKXBjGOQNrBGBTiaLOrYQqGZNPyd3xrQ4-xoB21QfXqrBbAVvtd9vrabcE3x_hqN2g9hP8R_8CuTtvQw</recordid><startdate>202110</startdate><enddate>202110</enddate><creator>Venezia, Jonathan H.</creator><creator>Whittle, Nicole</creator><creator>Herrera Ortiz, Christian</creator><creator>Leek, Marjorie R.</creator><creator>Barcenas, Caleb</creator><creator>Lee, Grace</creator><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>202110</creationdate><title>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</title><author>Venezia, Jonathan H. ; Whittle, Nicole ; Herrera Ortiz, Christian ; Leek, Marjorie R. ; Barcenas, Caleb ; Lee, Grace</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c715-570190dbfe6e3e1401c1046a1b87b6c28256635d02201fd51eb85493f618854d3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Venezia, Jonathan H.</creatorcontrib><creatorcontrib>Whittle, Nicole</creatorcontrib><creatorcontrib>Herrera Ortiz, Christian</creatorcontrib><creatorcontrib>Leek, Marjorie R.</creatorcontrib><creatorcontrib>Barcenas, Caleb</creatorcontrib><creatorcontrib>Lee, Grace</creatorcontrib><collection>CrossRef</collection><jtitle>The Journal of the Acoustical Society of America</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Venezia, Jonathan H.</au><au>Whittle, Nicole</au><au>Herrera Ortiz, Christian</au><au>Leek, Marjorie R.</au><au>Barcenas, Caleb</au><au>Lee, Grace</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</atitle><jtitle>The Journal of the Acoustical Society of America</jtitle><date>2021-10</date><risdate>2021</risdate><volume>150</volume><issue>4</issue><spage>A305</spage><epage>A305</epage><pages>A305-A305</pages><issn>0001-4966</issn><eissn>1520-8524</eissn><coden>JASMAN</coden><abstract>Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.</abstract><doi>10.1121/10.0008375</doi><tpages>1</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0001-4966
ispartof The Journal of the Acoustical Society of America, 2021-10, Vol.150 (4), p.A305-A305
issn 0001-4966
1520-8524
language eng
recordid cdi_crossref_primary_10_1121_10_0008375
source AIP Journals Complete; Alma/SFX Local Collection; AIP Acoustical Society of America
title Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T20%3A30%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-scitation_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Prediction%20of%20speech%20recognition%20in%20background%20noise%20and%20competing%20speech%20from%20suprathreshold%20auditory%20and%20cognitive%20measures&rft.jtitle=The%20Journal%20of%20the%20Acoustical%20Society%20of%20America&rft.au=Venezia,%20Jonathan%20H.&rft.date=2021-10&rft.volume=150&rft.issue=4&rft.spage=A305&rft.epage=A305&rft.pages=A305-A305&rft.issn=0001-4966&rft.eissn=1520-8524&rft.coden=JASMAN&rft_id=info:doi/10.1121/10.0008375&rft_dat=%3Cscitation_cross%3Ejasa%3C/scitation_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true