Speech synthesis method and device, electronic equipment and storage medium

The invention provides a speech synthesis method and device, electronic equipment and a storage medium, and the method comprises the steps: obtaining a to-be-synthesized text, and controlling the target emotion intensity of speech synthesis; determining a global emotion feature under the target emot...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: FANG XIN, WANG JINWEI, HU YAJUN, WU HONGCHUAN
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator FANG XIN
WANG JINWEI
HU YAJUN
WU HONGCHUAN
description The invention provides a speech synthesis method and device, electronic equipment and a storage medium, and the method comprises the steps: obtaining a to-be-synthesized text, and controlling the target emotion intensity of speech synthesis; determining a global emotion feature under the target emotion intensity based on the gear emotion feature of each preset emotion gear; speech synthesis is carried out based on the text and the global emotion characteristics, the defect that the emotion expression effect of the synthesized speech is poor due to the fact that the synthesized speech lacks emotions or is single in emotion expression and lacks conversion between emotions in a traditional scheme is overcome, speech synthesis is controlled through the global emotion characteristics, and the speech synthesis efficiency is improved. According to the method, the emotion direction and the emotion intensity can be positioned in the emotion space, information expression of the synthesized speech on different emotions
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN116013247A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN116013247A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN116013247A3</originalsourceid><addsrcrecordid>eNqNy70KwjAUhuEsDqLew3FXMFZ0lqIIgovuJSRfTaD5sedU8O4V8QKc3uV5x-p8LYD1xK8kHhyYIsRnRyY5cngGiwWhg5U-p2AJjyGUiCRfwJJ7c8fncWGIUzVqTceY_TpR8-PhVp-WKLkBF2ORIE190Xq70tV6s9tX_5g3atw1fw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Speech synthesis method and device, electronic equipment and storage medium</title><source>esp@cenet</source><creator>FANG XIN ; WANG JINWEI ; HU YAJUN ; WU HONGCHUAN</creator><creatorcontrib>FANG XIN ; WANG JINWEI ; HU YAJUN ; WU HONGCHUAN</creatorcontrib><description>The invention provides a speech synthesis method and device, electronic equipment and a storage medium, and the method comprises the steps: obtaining a to-be-synthesized text, and controlling the target emotion intensity of speech synthesis; determining a global emotion feature under the target emotion intensity based on the gear emotion feature of each preset emotion gear; speech synthesis is carried out based on the text and the global emotion characteristics, the defect that the emotion expression effect of the synthesized speech is poor due to the fact that the synthesized speech lacks emotions or is single in emotion expression and lacks conversion between emotions in a traditional scheme is overcome, speech synthesis is controlled through the global emotion characteristics, and the speech synthesis efficiency is improved. According to the method, the emotion direction and the emotion intensity can be positioned in the emotion space, information expression of the synthesized speech on different emotions</description><language>chi ; eng</language><subject>ACOUSTICS ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2023</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20230425&amp;DB=EPODOC&amp;CC=CN&amp;NR=116013247A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,777,882,25545,76296</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20230425&amp;DB=EPODOC&amp;CC=CN&amp;NR=116013247A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>FANG XIN</creatorcontrib><creatorcontrib>WANG JINWEI</creatorcontrib><creatorcontrib>HU YAJUN</creatorcontrib><creatorcontrib>WU HONGCHUAN</creatorcontrib><title>Speech synthesis method and device, electronic equipment and storage medium</title><description>The invention provides a speech synthesis method and device, electronic equipment and a storage medium, and the method comprises the steps: obtaining a to-be-synthesized text, and controlling the target emotion intensity of speech synthesis; determining a global emotion feature under the target emotion intensity based on the gear emotion feature of each preset emotion gear; speech synthesis is carried out based on the text and the global emotion characteristics, the defect that the emotion expression effect of the synthesized speech is poor due to the fact that the synthesized speech lacks emotions or is single in emotion expression and lacks conversion between emotions in a traditional scheme is overcome, speech synthesis is controlled through the global emotion characteristics, and the speech synthesis efficiency is improved. According to the method, the emotion direction and the emotion intensity can be positioned in the emotion space, information expression of the synthesized speech on different emotions</description><subject>ACOUSTICS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2023</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNy70KwjAUhuEsDqLew3FXMFZ0lqIIgovuJSRfTaD5sedU8O4V8QKc3uV5x-p8LYD1xK8kHhyYIsRnRyY5cngGiwWhg5U-p2AJjyGUiCRfwJJ7c8fncWGIUzVqTceY_TpR8-PhVp-WKLkBF2ORIE190Xq70tV6s9tX_5g3atw1fw</recordid><startdate>20230425</startdate><enddate>20230425</enddate><creator>FANG XIN</creator><creator>WANG JINWEI</creator><creator>HU YAJUN</creator><creator>WU HONGCHUAN</creator><scope>EVB</scope></search><sort><creationdate>20230425</creationdate><title>Speech synthesis method and device, electronic equipment and storage medium</title><author>FANG XIN ; WANG JINWEI ; HU YAJUN ; WU HONGCHUAN</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN116013247A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2023</creationdate><topic>ACOUSTICS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>FANG XIN</creatorcontrib><creatorcontrib>WANG JINWEI</creatorcontrib><creatorcontrib>HU YAJUN</creatorcontrib><creatorcontrib>WU HONGCHUAN</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>FANG XIN</au><au>WANG JINWEI</au><au>HU YAJUN</au><au>WU HONGCHUAN</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Speech synthesis method and device, electronic equipment and storage medium</title><date>2023-04-25</date><risdate>2023</risdate><abstract>The invention provides a speech synthesis method and device, electronic equipment and a storage medium, and the method comprises the steps: obtaining a to-be-synthesized text, and controlling the target emotion intensity of speech synthesis; determining a global emotion feature under the target emotion intensity based on the gear emotion feature of each preset emotion gear; speech synthesis is carried out based on the text and the global emotion characteristics, the defect that the emotion expression effect of the synthesized speech is poor due to the fact that the synthesized speech lacks emotions or is single in emotion expression and lacks conversion between emotions in a traditional scheme is overcome, speech synthesis is controlled through the global emotion characteristics, and the speech synthesis efficiency is improved. According to the method, the emotion direction and the emotion intensity can be positioned in the emotion space, information expression of the synthesized speech on different emotions</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN116013247A
source esp@cenet
subjects ACOUSTICS
MUSICAL INSTRUMENTS
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title Speech synthesis method and device, electronic equipment and storage medium
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T01%3A58%3A08IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=FANG%20XIN&rft.date=2023-04-25&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN116013247A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true