The contribution of phonation type to the perception of vocal emotions in German: an articulatory synthesis study
Vocal emotions are signaled by specific patterns of prosodic parameters, most notably pitch, phone duration, intensity, and phonation type. Phonation type was so far the least accessible parameter in emotion research, because it was difficult to extract from speech signals and difficult to manipulat...
Gespeichert in:
Veröffentlicht in: | The Journal of the Acoustical Society of America 2015-03, Vol.137 (3), p.1503-1512 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1512 |
---|---|
container_issue | 3 |
container_start_page | 1503 |
container_title | The Journal of the Acoustical Society of America |
container_volume | 137 |
creator | Birkholz, Peter Martin, Lucia Willmes, Klaus Kröger, Bernd J Neuschaefer-Rube, Christiane |
description | Vocal emotions are signaled by specific patterns of prosodic parameters, most notably pitch, phone duration, intensity, and phonation type. Phonation type was so far the least accessible parameter in emotion research, because it was difficult to extract from speech signals and difficult to manipulate in natural or synthetic speech. The present study built on recent advances in articulatory speech synthesis to exclusively control phonation type in re-synthesized German sentences spoken with seven different emotions. The goal was to find out to what extent the sole change of phonation type affects the perception of these emotions. Therefore, portrayed emotional utterances were re-synthesized with their original phonation type, as well as with each purely breathy, modal, and pressed phonation, and then rated by listeners with respect to the perceived emotions. Highly significant effects of phonation type on the recognition rates of the original emotions were found, except for disgust. While fear, anger, and the neutral emotion require specific phonation types for correct perception, sadness, happiness, boredom, and disgust primarily rely on other prosodic parameters. These results can help to improve the expression of emotions in synthesized speech and facilitate the robust automatic recognition of vocal emotions. |
doi_str_mv | 10.1121/1.4906836 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1665122562</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1665122562</sourcerecordid><originalsourceid>FETCH-LOGICAL-c351t-70e167b1af083c4cbb1b061e742040cd370731454f0f55d833007d31c1f7584b3</originalsourceid><addsrcrecordid>eNo9kD1PwzAURS0EoqUw8AeQRxhS_PyVhA1VUJAqsZQ5chxHDUrs1HaQ8u9JoWV6uk_n3uEgdAtkCUDhEZY8JzJj8gzNQVCSZILyczQnhEDCcyln6CqErymKjOWXaEZFmslcwhzttzuDtbPRN-UQG2exq3G_c1b9hjj2BkeH40T1xmvTn5hvp1WLTecOj4Abi9fGd8o-YWWx8rHRQ6ui8yMOo53qoQk4xKEar9FFrdpgbo53gT5fX7art2TzsX5fPW8SzQTEJCUGZFqCqknGNNdlCSWRYFJOCSe6YilJGXDBa1ILUWWMEZJWDDTUqch4yRbo_m-3924_mBCLrgnatK2yxg2hACkFUCokndCHP1R7F4I3ddH7plN-LIAUB8MFFEfDE3t3nB3KzlT_5Ekp-wE_jXYo</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1665122562</pqid></control><display><type>article</type><title>The contribution of phonation type to the perception of vocal emotions in German: an articulatory synthesis study</title><source>MEDLINE</source><source>AIP Journals Complete</source><source>Alma/SFX Local Collection</source><source>AIP Acoustical Society of America</source><creator>Birkholz, Peter ; Martin, Lucia ; Willmes, Klaus ; Kröger, Bernd J ; Neuschaefer-Rube, Christiane</creator><creatorcontrib>Birkholz, Peter ; Martin, Lucia ; Willmes, Klaus ; Kröger, Bernd J ; Neuschaefer-Rube, Christiane</creatorcontrib><description>Vocal emotions are signaled by specific patterns of prosodic parameters, most notably pitch, phone duration, intensity, and phonation type. Phonation type was so far the least accessible parameter in emotion research, because it was difficult to extract from speech signals and difficult to manipulate in natural or synthetic speech. The present study built on recent advances in articulatory speech synthesis to exclusively control phonation type in re-synthesized German sentences spoken with seven different emotions. The goal was to find out to what extent the sole change of phonation type affects the perception of these emotions. Therefore, portrayed emotional utterances were re-synthesized with their original phonation type, as well as with each purely breathy, modal, and pressed phonation, and then rated by listeners with respect to the perceived emotions. Highly significant effects of phonation type on the recognition rates of the original emotions were found, except for disgust. While fear, anger, and the neutral emotion require specific phonation types for correct perception, sadness, happiness, boredom, and disgust primarily rely on other prosodic parameters. These results can help to improve the expression of emotions in synthesized speech and facilitate the robust automatic recognition of vocal emotions.</description><identifier>ISSN: 0001-4966</identifier><identifier>EISSN: 1520-8524</identifier><identifier>DOI: 10.1121/1.4906836</identifier><identifier>PMID: 25786961</identifier><language>eng</language><publisher>United States</publisher><subject>Acoustics ; Adult ; Emotions ; Female ; Humans ; Loudness Perception ; Male ; Pattern Recognition, Automated ; Phonation ; Phonetics ; Pitch Perception ; Signal Processing, Computer-Assisted ; Sound Spectrography ; Speech Acoustics ; Speech Perception ; Speech Production Measurement ; Time Factors ; Voice Quality ; Young Adult</subject><ispartof>The Journal of the Acoustical Society of America, 2015-03, Vol.137 (3), p.1503-1512</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c351t-70e167b1af083c4cbb1b061e742040cd370731454f0f55d833007d31c1f7584b3</citedby><cites>FETCH-LOGICAL-c351t-70e167b1af083c4cbb1b061e742040cd370731454f0f55d833007d31c1f7584b3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>207,208,314,777,781,27905,27906</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/25786961$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Birkholz, Peter</creatorcontrib><creatorcontrib>Martin, Lucia</creatorcontrib><creatorcontrib>Willmes, Klaus</creatorcontrib><creatorcontrib>Kröger, Bernd J</creatorcontrib><creatorcontrib>Neuschaefer-Rube, Christiane</creatorcontrib><title>The contribution of phonation type to the perception of vocal emotions in German: an articulatory synthesis study</title><title>The Journal of the Acoustical Society of America</title><addtitle>J Acoust Soc Am</addtitle><description>Vocal emotions are signaled by specific patterns of prosodic parameters, most notably pitch, phone duration, intensity, and phonation type. Phonation type was so far the least accessible parameter in emotion research, because it was difficult to extract from speech signals and difficult to manipulate in natural or synthetic speech. The present study built on recent advances in articulatory speech synthesis to exclusively control phonation type in re-synthesized German sentences spoken with seven different emotions. The goal was to find out to what extent the sole change of phonation type affects the perception of these emotions. Therefore, portrayed emotional utterances were re-synthesized with their original phonation type, as well as with each purely breathy, modal, and pressed phonation, and then rated by listeners with respect to the perceived emotions. Highly significant effects of phonation type on the recognition rates of the original emotions were found, except for disgust. While fear, anger, and the neutral emotion require specific phonation types for correct perception, sadness, happiness, boredom, and disgust primarily rely on other prosodic parameters. These results can help to improve the expression of emotions in synthesized speech and facilitate the robust automatic recognition of vocal emotions.</description><subject>Acoustics</subject><subject>Adult</subject><subject>Emotions</subject><subject>Female</subject><subject>Humans</subject><subject>Loudness Perception</subject><subject>Male</subject><subject>Pattern Recognition, Automated</subject><subject>Phonation</subject><subject>Phonetics</subject><subject>Pitch Perception</subject><subject>Signal Processing, Computer-Assisted</subject><subject>Sound Spectrography</subject><subject>Speech Acoustics</subject><subject>Speech Perception</subject><subject>Speech Production Measurement</subject><subject>Time Factors</subject><subject>Voice Quality</subject><subject>Young Adult</subject><issn>0001-4966</issn><issn>1520-8524</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2015</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNo9kD1PwzAURS0EoqUw8AeQRxhS_PyVhA1VUJAqsZQ5chxHDUrs1HaQ8u9JoWV6uk_n3uEgdAtkCUDhEZY8JzJj8gzNQVCSZILyczQnhEDCcyln6CqErymKjOWXaEZFmslcwhzttzuDtbPRN-UQG2exq3G_c1b9hjj2BkeH40T1xmvTn5hvp1WLTecOj4Abi9fGd8o-YWWx8rHRQ6ui8yMOo53qoQk4xKEar9FFrdpgbo53gT5fX7art2TzsX5fPW8SzQTEJCUGZFqCqknGNNdlCSWRYFJOCSe6YilJGXDBa1ILUWWMEZJWDDTUqch4yRbo_m-3924_mBCLrgnatK2yxg2hACkFUCokndCHP1R7F4I3ddH7plN-LIAUB8MFFEfDE3t3nB3KzlT_5Ekp-wE_jXYo</recordid><startdate>201503</startdate><enddate>201503</enddate><creator>Birkholz, Peter</creator><creator>Martin, Lucia</creator><creator>Willmes, Klaus</creator><creator>Kröger, Bernd J</creator><creator>Neuschaefer-Rube, Christiane</creator><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope></search><sort><creationdate>201503</creationdate><title>The contribution of phonation type to the perception of vocal emotions in German: an articulatory synthesis study</title><author>Birkholz, Peter ; Martin, Lucia ; Willmes, Klaus ; Kröger, Bernd J ; Neuschaefer-Rube, Christiane</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c351t-70e167b1af083c4cbb1b061e742040cd370731454f0f55d833007d31c1f7584b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2015</creationdate><topic>Acoustics</topic><topic>Adult</topic><topic>Emotions</topic><topic>Female</topic><topic>Humans</topic><topic>Loudness Perception</topic><topic>Male</topic><topic>Pattern Recognition, Automated</topic><topic>Phonation</topic><topic>Phonetics</topic><topic>Pitch Perception</topic><topic>Signal Processing, Computer-Assisted</topic><topic>Sound Spectrography</topic><topic>Speech Acoustics</topic><topic>Speech Perception</topic><topic>Speech Production Measurement</topic><topic>Time Factors</topic><topic>Voice Quality</topic><topic>Young Adult</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Birkholz, Peter</creatorcontrib><creatorcontrib>Martin, Lucia</creatorcontrib><creatorcontrib>Willmes, Klaus</creatorcontrib><creatorcontrib>Kröger, Bernd J</creatorcontrib><creatorcontrib>Neuschaefer-Rube, Christiane</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>The Journal of the Acoustical Society of America</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Birkholz, Peter</au><au>Martin, Lucia</au><au>Willmes, Klaus</au><au>Kröger, Bernd J</au><au>Neuschaefer-Rube, Christiane</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>The contribution of phonation type to the perception of vocal emotions in German: an articulatory synthesis study</atitle><jtitle>The Journal of the Acoustical Society of America</jtitle><addtitle>J Acoust Soc Am</addtitle><date>2015-03</date><risdate>2015</risdate><volume>137</volume><issue>3</issue><spage>1503</spage><epage>1512</epage><pages>1503-1512</pages><issn>0001-4966</issn><eissn>1520-8524</eissn><abstract>Vocal emotions are signaled by specific patterns of prosodic parameters, most notably pitch, phone duration, intensity, and phonation type. Phonation type was so far the least accessible parameter in emotion research, because it was difficult to extract from speech signals and difficult to manipulate in natural or synthetic speech. The present study built on recent advances in articulatory speech synthesis to exclusively control phonation type in re-synthesized German sentences spoken with seven different emotions. The goal was to find out to what extent the sole change of phonation type affects the perception of these emotions. Therefore, portrayed emotional utterances were re-synthesized with their original phonation type, as well as with each purely breathy, modal, and pressed phonation, and then rated by listeners with respect to the perceived emotions. Highly significant effects of phonation type on the recognition rates of the original emotions were found, except for disgust. While fear, anger, and the neutral emotion require specific phonation types for correct perception, sadness, happiness, boredom, and disgust primarily rely on other prosodic parameters. These results can help to improve the expression of emotions in synthesized speech and facilitate the robust automatic recognition of vocal emotions.</abstract><cop>United States</cop><pmid>25786961</pmid><doi>10.1121/1.4906836</doi><tpages>10</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0001-4966 |
ispartof | The Journal of the Acoustical Society of America, 2015-03, Vol.137 (3), p.1503-1512 |
issn | 0001-4966 1520-8524 |
language | eng |
recordid | cdi_proquest_miscellaneous_1665122562 |
source | MEDLINE; AIP Journals Complete; Alma/SFX Local Collection; AIP Acoustical Society of America |
subjects | Acoustics Adult Emotions Female Humans Loudness Perception Male Pattern Recognition, Automated Phonation Phonetics Pitch Perception Signal Processing, Computer-Assisted Sound Spectrography Speech Acoustics Speech Perception Speech Production Measurement Time Factors Voice Quality Young Adult |
title | The contribution of phonation type to the perception of vocal emotions in German: an articulatory synthesis study |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T04%3A49%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=The%20contribution%20of%20phonation%20type%20to%20the%20perception%20of%20vocal%20emotions%20in%20German:%20an%20articulatory%20synthesis%20study&rft.jtitle=The%20Journal%20of%20the%20Acoustical%20Society%20of%20America&rft.au=Birkholz,%20Peter&rft.date=2015-03&rft.volume=137&rft.issue=3&rft.spage=1503&rft.epage=1512&rft.pages=1503-1512&rft.issn=0001-4966&rft.eissn=1520-8524&rft_id=info:doi/10.1121/1.4906836&rft_dat=%3Cproquest_cross%3E1665122562%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1665122562&rft_id=info:pmid/25786961&rfr_iscdi=true |