Method and device for dynamically extracting speech emotion features
The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are fir...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Patent |
Sprache: | chi ; eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | XIA WEI LIU RUQIAN ZHONG HONGMEI DOU SHUWEI HAN TINGTING |
description | The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot |
format | Patent |
fullrecord | <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN117373491A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN117373491A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN117373491A3</originalsourceid><addsrcrecordid>eNqNyjEOAiEQBVAaC6PeYTyABcFkY2lWjY1W9psJfFwSFgiMxr29jQewes1bqtMNMmZHnBw5vIMF-VzJzYmnYDnGmfCRylZCelIrgB0JU5aQE3mwvCraWi08x4bNz5XaXs6P_rpDyQNaYYsEGfq71p3pzP6gj-af8wXL2DMg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Method and device for dynamically extracting speech emotion features</title><source>esp@cenet</source><creator>XIA WEI ; LIU RUQIAN ; ZHONG HONGMEI ; DOU SHUWEI ; HAN TINGTING</creator><creatorcontrib>XIA WEI ; LIU RUQIAN ; ZHONG HONGMEI ; DOU SHUWEI ; HAN TINGTING</creatorcontrib><description>The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot</description><language>chi ; eng</language><subject>ACOUSTICS ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20240109&DB=EPODOC&CC=CN&NR=117373491A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20240109&DB=EPODOC&CC=CN&NR=117373491A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>XIA WEI</creatorcontrib><creatorcontrib>LIU RUQIAN</creatorcontrib><creatorcontrib>ZHONG HONGMEI</creatorcontrib><creatorcontrib>DOU SHUWEI</creatorcontrib><creatorcontrib>HAN TINGTING</creatorcontrib><title>Method and device for dynamically extracting speech emotion features</title><description>The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot</description><subject>ACOUSTICS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNyjEOAiEQBVAaC6PeYTyABcFkY2lWjY1W9psJfFwSFgiMxr29jQewes1bqtMNMmZHnBw5vIMF-VzJzYmnYDnGmfCRylZCelIrgB0JU5aQE3mwvCraWi08x4bNz5XaXs6P_rpDyQNaYYsEGfq71p3pzP6gj-af8wXL2DMg</recordid><startdate>20240109</startdate><enddate>20240109</enddate><creator>XIA WEI</creator><creator>LIU RUQIAN</creator><creator>ZHONG HONGMEI</creator><creator>DOU SHUWEI</creator><creator>HAN TINGTING</creator><scope>EVB</scope></search><sort><creationdate>20240109</creationdate><title>Method and device for dynamically extracting speech emotion features</title><author>XIA WEI ; LIU RUQIAN ; ZHONG HONGMEI ; DOU SHUWEI ; HAN TINGTING</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN117373491A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2024</creationdate><topic>ACOUSTICS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>XIA WEI</creatorcontrib><creatorcontrib>LIU RUQIAN</creatorcontrib><creatorcontrib>ZHONG HONGMEI</creatorcontrib><creatorcontrib>DOU SHUWEI</creatorcontrib><creatorcontrib>HAN TINGTING</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>XIA WEI</au><au>LIU RUQIAN</au><au>ZHONG HONGMEI</au><au>DOU SHUWEI</au><au>HAN TINGTING</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Method and device for dynamically extracting speech emotion features</title><date>2024-01-09</date><risdate>2024</risdate><abstract>The invention discloses a method and a device for dynamically extracting speech emotion characteristics, which are characterized in that forward input data and reverse input data of speech data are respectively imported into N frame-level characteristic encoders, and the imported speech data are firstly used for dynamically extracting the frame-level fusion characteristics of the speech through a frame-level dynamic fusion unit; then cross-scale information between frames is obtained through a one-dimensional time sequence convolution unit, after normalization and activation processing, attention weight distribution is conducted on the obtained information through an attention unit, and the information acts on imported voice data. And respectively outputting the N forward voice emotion features and the N reverse voice emotion features, and importing the N forward voice emotion features and the N reverse voice emotion features into a global feature encoder for emotion fusion to obtain final advanced voice emot</abstract><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | |
ispartof | |
issn | |
language | chi ; eng |
recordid | cdi_epo_espacenet_CN117373491A |
source | esp@cenet |
subjects | ACOUSTICS MUSICAL INSTRUMENTS PHYSICS SPEECH ANALYSIS OR SYNTHESIS SPEECH OR AUDIO CODING OR DECODING SPEECH OR VOICE PROCESSING SPEECH RECOGNITION |
title | Method and device for dynamically extracting speech emotion features |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T23%3A25%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=XIA%20WEI&rft.date=2024-01-09&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN117373491A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |