Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures
Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audio...
Gespeichert in:
Veröffentlicht in: | The Journal of the Acoustical Society of America 2021-10, Vol.150 (4), p.A305-A305 |
---|---|
Hauptverfasser: | , , , , , |
Format: | Artikel |
Sprache: | eng |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | A305 |
---|---|
container_issue | 4 |
container_start_page | A305 |
container_title | The Journal of the Acoustical Society of America |
container_volume | 150 |
creator | Venezia, Jonathan H. Whittle, Nicole Herrera Ortiz, Christian Leek, Marjorie R. Barcenas, Caleb Lee, Grace |
description | Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent. |
doi_str_mv | 10.1121/10.0008375 |
format | Article |
fullrecord | <record><control><sourceid>scitation_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1121_10_0008375</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>jasa</sourcerecordid><originalsourceid>FETCH-LOGICAL-c715-570190dbfe6e3e1401c1046a1b87b6c28256635d02201fd51eb85493f618854d3</originalsourceid><addsrcrecordid>eNp9kD9PwzAUxC0EEqWw8Ak8gwJ-TuwkI6qgIFWCoXvl2M-toYkjO0HqwHfH_cPKdPdOv3vDEXIL7AGAw2NSxliVl-KMTEBwllWCF-dkklLIilrKS3IV42c6RZXXE_LzEdA4PTjfUW9p7BH1hgbUft25Q-o62ij9tQ5-7AztvItIVXLatz0Orlv_lWzwLY1jH9SwCRg3fmuoGo0bfNidGoef30hbVHFMzDW5sGob8eakU7J8eV7OXrPF-_xt9rTIdAkiEyWDmpnGosQcoWCggRVSQVOVjdS84kLKXBjGOQNrBGBTiaLOrYQqGZNPyd3xrQ4-xoB21QfXqrBbAVvtd9vrabcE3x_hqN2g9hP8R_8CuTtvQw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</title><source>AIP Journals Complete</source><source>Alma/SFX Local Collection</source><source>AIP Acoustical Society of America</source><creator>Venezia, Jonathan H. ; Whittle, Nicole ; Herrera Ortiz, Christian ; Leek, Marjorie R. ; Barcenas, Caleb ; Lee, Grace</creator><creatorcontrib>Venezia, Jonathan H. ; Whittle, Nicole ; Herrera Ortiz, Christian ; Leek, Marjorie R. ; Barcenas, Caleb ; Lee, Grace</creatorcontrib><description>Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.</description><identifier>ISSN: 0001-4966</identifier><identifier>EISSN: 1520-8524</identifier><identifier>DOI: 10.1121/10.0008375</identifier><identifier>CODEN: JASMAN</identifier><language>eng</language><ispartof>The Journal of the Acoustical Society of America, 2021-10, Vol.150 (4), p.A305-A305</ispartof><rights>Acoustical Society of America</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://pubs.aip.org/jasa/article-lookup/doi/10.1121/10.0008375$$EHTML$$P50$$Gscitation$$H</linktohtml><link.rule.ids>207,208,314,776,780,790,1559,4498,27901,27902,76126</link.rule.ids></links><search><creatorcontrib>Venezia, Jonathan H.</creatorcontrib><creatorcontrib>Whittle, Nicole</creatorcontrib><creatorcontrib>Herrera Ortiz, Christian</creatorcontrib><creatorcontrib>Leek, Marjorie R.</creatorcontrib><creatorcontrib>Barcenas, Caleb</creatorcontrib><creatorcontrib>Lee, Grace</creatorcontrib><title>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</title><title>The Journal of the Acoustical Society of America</title><description>Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.</description><issn>0001-4966</issn><issn>1520-8524</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kD9PwzAUxC0EEqWw8Ak8gwJ-TuwkI6qgIFWCoXvl2M-toYkjO0HqwHfH_cPKdPdOv3vDEXIL7AGAw2NSxliVl-KMTEBwllWCF-dkklLIilrKS3IV42c6RZXXE_LzEdA4PTjfUW9p7BH1hgbUft25Q-o62ij9tQ5-7AztvItIVXLatz0Orlv_lWzwLY1jH9SwCRg3fmuoGo0bfNidGoef30hbVHFMzDW5sGob8eakU7J8eV7OXrPF-_xt9rTIdAkiEyWDmpnGosQcoWCggRVSQVOVjdS84kLKXBjGOQNrBGBTiaLOrYQqGZNPyd3xrQ4-xoB21QfXqrBbAVvtd9vrabcE3x_hqN2g9hP8R_8CuTtvQw</recordid><startdate>202110</startdate><enddate>202110</enddate><creator>Venezia, Jonathan H.</creator><creator>Whittle, Nicole</creator><creator>Herrera Ortiz, Christian</creator><creator>Leek, Marjorie R.</creator><creator>Barcenas, Caleb</creator><creator>Lee, Grace</creator><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>202110</creationdate><title>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</title><author>Venezia, Jonathan H. ; Whittle, Nicole ; Herrera Ortiz, Christian ; Leek, Marjorie R. ; Barcenas, Caleb ; Lee, Grace</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c715-570190dbfe6e3e1401c1046a1b87b6c28256635d02201fd51eb85493f618854d3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Venezia, Jonathan H.</creatorcontrib><creatorcontrib>Whittle, Nicole</creatorcontrib><creatorcontrib>Herrera Ortiz, Christian</creatorcontrib><creatorcontrib>Leek, Marjorie R.</creatorcontrib><creatorcontrib>Barcenas, Caleb</creatorcontrib><creatorcontrib>Lee, Grace</creatorcontrib><collection>CrossRef</collection><jtitle>The Journal of the Acoustical Society of America</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Venezia, Jonathan H.</au><au>Whittle, Nicole</au><au>Herrera Ortiz, Christian</au><au>Leek, Marjorie R.</au><au>Barcenas, Caleb</au><au>Lee, Grace</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures</atitle><jtitle>The Journal of the Acoustical Society of America</jtitle><date>2021-10</date><risdate>2021</risdate><volume>150</volume><issue>4</issue><spage>A305</spage><epage>A305</epage><pages>A305-A305</pages><issn>0001-4966</issn><eissn>1520-8524</eissn><coden>JASMAN</coden><abstract>Previous studies have struggled to identify measures beyond the audiogram to reliably predict speech-in-noise scores. This may owe to: (i) different mechanisms mediate performance depending on materials and task; and (ii) effects are not reproducible. Here, 38 listeners with normal/near-normal audiograms completed batteries of temporal auditory and cognitive tests, and speech recognition (“Theo-Victor-Michael” test) in speech-shaped noise (SSN), speech-envelope modulated noise (envSSN), one (1T) and two (2T) competing talkers. A two-stage Bayesian modeling approach was employed. In Stage 1, speech scores were corrected for target-word frequency/neighborhood density, psychometric function parameters were extracted from temporal tests, and cognitive measures were reduced to three composite variables. Stage 2 then applied Gaussian process models to predict speech scores from temporal and cognitive measures. Leave-one-out cross-validation and model stacking determined the best combination of predictive models. Performance in SSN/envSSN was best predicted by temporal envelope measures (forward masking, gap duration discrimination), while performance in 1T was best predicted by cognitive measures (executive function, processing speed). Temporal fine structure measures (frequency-modulation, interaural-phase-difference detection) predicted the number of 1T distractor responses. All models failed on 2T. These results show that prediction of speech-in-noise scores from suprathreshold “process” measures is highly task dependent.</abstract><doi>10.1121/10.0008375</doi><tpages>1</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0001-4966 |
ispartof | The Journal of the Acoustical Society of America, 2021-10, Vol.150 (4), p.A305-A305 |
issn | 0001-4966 1520-8524 |
language | eng |
recordid | cdi_crossref_primary_10_1121_10_0008375 |
source | AIP Journals Complete; Alma/SFX Local Collection; AIP Acoustical Society of America |
title | Prediction of speech recognition in background noise and competing speech from suprathreshold auditory and cognitive measures |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T20%3A30%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-scitation_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Prediction%20of%20speech%20recognition%20in%20background%20noise%20and%20competing%20speech%20from%20suprathreshold%20auditory%20and%20cognitive%20measures&rft.jtitle=The%20Journal%20of%20the%20Acoustical%20Society%20of%20America&rft.au=Venezia,%20Jonathan%20H.&rft.date=2021-10&rft.volume=150&rft.issue=4&rft.spage=A305&rft.epage=A305&rft.pages=A305-A305&rft.issn=0001-4966&rft.eissn=1520-8524&rft.coden=JASMAN&rft_id=info:doi/10.1121/10.0008375&rft_dat=%3Cscitation_cross%3Ejasa%3C/scitation_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |