Method and device for dynamically extracting speech emotion features

The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are fir...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: XIA WEI, LIU RUQIAN, ZHONG HONGMEI, DOU SHUWEI, HAN TINGTING
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator XIA WEI
LIU RUQIAN
ZHONG HONGMEI
DOU SHUWEI
HAN TINGTING
description The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN117373491A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN117373491A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN117373491A3</originalsourceid><addsrcrecordid>eNqNyjEOAiEQBVAaC6PeYTyABcFkY2lWjY1W9psJfFwSFgiMxr29jQewes1bqtMNMmZHnBw5vIMF-VzJzYmnYDnGmfCRylZCelIrgB0JU5aQE3mwvCraWi08x4bNz5XaXs6P_rpDyQNaYYsEGfq71p3pzP6gj-af8wXL2DMg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Method and device for dynamically extracting speech emotion features</title><source>esp@cenet</source><creator>XIA WEI ; LIU RUQIAN ; ZHONG HONGMEI ; DOU SHUWEI ; HAN TINGTING</creator><creatorcontrib>XIA WEI ; LIU RUQIAN ; ZHONG HONGMEI ; DOU SHUWEI ; HAN TINGTING</creatorcontrib><description>The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot</description><language>chi ; eng</language><subject>ACOUSTICS ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240109&amp;DB=EPODOC&amp;CC=CN&amp;NR=117373491A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240109&amp;DB=EPODOC&amp;CC=CN&amp;NR=117373491A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>XIA WEI</creatorcontrib><creatorcontrib>LIU RUQIAN</creatorcontrib><creatorcontrib>ZHONG HONGMEI</creatorcontrib><creatorcontrib>DOU SHUWEI</creatorcontrib><creatorcontrib>HAN TINGTING</creatorcontrib><title>Method and device for dynamically extracting speech emotion features</title><description>The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot</description><subject>ACOUSTICS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNyjEOAiEQBVAaC6PeYTyABcFkY2lWjY1W9psJfFwSFgiMxr29jQewes1bqtMNMmZHnBw5vIMF-VzJzYmnYDnGmfCRylZCelIrgB0JU5aQE3mwvCraWi08x4bNz5XaXs6P_rpDyQNaYYsEGfq71p3pzP6gj-af8wXL2DMg</recordid><startdate>20240109</startdate><enddate>20240109</enddate><creator>XIA WEI</creator><creator>LIU RUQIAN</creator><creator>ZHONG HONGMEI</creator><creator>DOU SHUWEI</creator><creator>HAN TINGTING</creator><scope>EVB</scope></search><sort><creationdate>20240109</creationdate><title>Method and device for dynamically extracting speech emotion features</title><author>XIA WEI ; LIU RUQIAN ; ZHONG HONGMEI ; DOU SHUWEI ; HAN TINGTING</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN117373491A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2024</creationdate><topic>ACOUSTICS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>XIA WEI</creatorcontrib><creatorcontrib>LIU RUQIAN</creatorcontrib><creatorcontrib>ZHONG HONGMEI</creatorcontrib><creatorcontrib>DOU SHUWEI</creatorcontrib><creatorcontrib>HAN TINGTING</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>XIA WEI</au><au>LIU RUQIAN</au><au>ZHONG HONGMEI</au><au>DOU SHUWEI</au><au>HAN TINGTING</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Method and device for dynamically extracting speech emotion features</title><date>2024-01-09</date><risdate>2024</risdate><abstract>The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN117373491A
source esp@cenet
subjects ACOUSTICS
MUSICAL INSTRUMENTS
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title Method and device for dynamically extracting speech emotion features
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T23%3A25%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=XIA%20WEI&rft.date=2024-01-09&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN117373491A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true