Speech synthesis method, device, equipment and medium

The invention discloses a speech synthesis method and device, equipment and a medium, and relates to the field of big data or finance. The method comprises the following steps: acquiring a voice data set; preprocessing the voice data set to obtain a training data set; training a speech synthesis mod...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: YAO HONGYAN, SU FANGXING, JI CHENG
Format: Patent
Sprache:chi ; eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator YAO HONGYAN
SU FANGXING
JI CHENG
description The invention discloses a speech synthesis method and device, equipment and a medium, and relates to the field of big data or finance. The method comprises the following steps: acquiring a voice data set; preprocessing the voice data set to obtain a training data set; training a speech synthesis model according to the training data set; obtaining synthetic data; and inputting the synthesis data into the speech synthesis model to perform speech synthesis. Therefore, the wave network model or the Tarkateon network model is trained through the training data set, and the speech synthesis model can be generated, so that the speech generation model can synthesize speech with high speech naturalness and emotional degree, and the speech synthesis experience of a user is improved. 本申请公开了一种语音合成方法、装置、设备及介质,涉及大数据领域或金融领域。该方法包括:获取语音数据集;对语音数据集进行预处理,以获取训练数据集;根据训练数据集,训练语音合成模型;获取合成数据;将合成数据输入至语音合成模型中,以执行语音合成。由此,通过训练数据集训练波网网络模型或塔科特龙网络模型,可以生成语音合成模型,使得该语音生成模型能够合成语音自然度、情感度高的语音,提升用户的语音合成体验。
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN116665640A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN116665640A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN116665640A3</originalsourceid><addsrcrecordid>eNrjZDANLkhNTc5QKK7MK8lILc4sVshNLcnIT9FRSEkty0xO1VFILSzNLMhNzStRSMxLAcqmZJbm8jCwpiXmFKfyQmluBkU31xBnD93Ugvz41OKCxOTUvNSSeGc_Q0MzMzNTMxMDR2Ni1AAAUvEs-Q</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Speech synthesis method, device, equipment and medium</title><source>esp@cenet</source><creator>YAO HONGYAN ; SU FANGXING ; JI CHENG</creator><creatorcontrib>YAO HONGYAN ; SU FANGXING ; JI CHENG</creatorcontrib><description>The invention discloses a speech synthesis method and device, equipment and a medium, and relates to the field of big data or finance. The method comprises the following steps: acquiring a voice data set; preprocessing the voice data set to obtain a training data set; training a speech synthesis model according to the training data set; obtaining synthetic data; and inputting the synthesis data into the speech synthesis model to perform speech synthesis. Therefore, the wave network model or the Tarkateon network model is trained through the training data set, and the speech synthesis model can be generated, so that the speech generation model can synthesize speech with high speech naturalness and emotional degree, and the speech synthesis experience of a user is improved. 本申请公开了一种语音合成方法、装置、设备及介质,涉及大数据领域或金融领域。该方法包括:获取语音数据集;对语音数据集进行预处理,以获取训练数据集;根据训练数据集,训练语音合成模型;获取合成数据;将合成数据输入至语音合成模型中,以执行语音合成。由此,通过训练数据集训练波网网络模型或塔科特龙网络模型,可以生成语音合成模型,使得该语音生成模型能够合成语音自然度、情感度高的语音,提升用户的语音合成体验。</description><language>chi ; eng</language><subject>ACOUSTICS ; CALCULATING ; COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS ; COMPUTING ; COUNTING ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2023</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20230829&amp;DB=EPODOC&amp;CC=CN&amp;NR=116665640A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,777,882,25546,76297</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20230829&amp;DB=EPODOC&amp;CC=CN&amp;NR=116665640A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>YAO HONGYAN</creatorcontrib><creatorcontrib>SU FANGXING</creatorcontrib><creatorcontrib>JI CHENG</creatorcontrib><title>Speech synthesis method, device, equipment and medium</title><description>The invention discloses a speech synthesis method and device, equipment and a medium, and relates to the field of big data or finance. The method comprises the following steps: acquiring a voice data set; preprocessing the voice data set to obtain a training data set; training a speech synthesis model according to the training data set; obtaining synthetic data; and inputting the synthesis data into the speech synthesis model to perform speech synthesis. Therefore, the wave network model or the Tarkateon network model is trained through the training data set, and the speech synthesis model can be generated, so that the speech generation model can synthesize speech with high speech naturalness and emotional degree, and the speech synthesis experience of a user is improved. 本申请公开了一种语音合成方法、装置、设备及介质,涉及大数据领域或金融领域。该方法包括:获取语音数据集;对语音数据集进行预处理,以获取训练数据集;根据训练数据集,训练语音合成模型;获取合成数据;将合成数据输入至语音合成模型中,以执行语音合成。由此,通过训练数据集训练波网网络模型或塔科特龙网络模型,可以生成语音合成模型,使得该语音生成模型能够合成语音自然度、情感度高的语音,提升用户的语音合成体验。</description><subject>ACOUSTICS</subject><subject>CALCULATING</subject><subject>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2023</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZDANLkhNTc5QKK7MK8lILc4sVshNLcnIT9FRSEkty0xO1VFILSzNLMhNzStRSMxLAcqmZJbm8jCwpiXmFKfyQmluBkU31xBnD93Ugvz41OKCxOTUvNSSeGc_Q0MzMzNTMxMDR2Ni1AAAUvEs-Q</recordid><startdate>20230829</startdate><enddate>20230829</enddate><creator>YAO HONGYAN</creator><creator>SU FANGXING</creator><creator>JI CHENG</creator><scope>EVB</scope></search><sort><creationdate>20230829</creationdate><title>Speech synthesis method, device, equipment and medium</title><author>YAO HONGYAN ; SU FANGXING ; JI CHENG</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN116665640A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2023</creationdate><topic>ACOUSTICS</topic><topic>CALCULATING</topic><topic>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>YAO HONGYAN</creatorcontrib><creatorcontrib>SU FANGXING</creatorcontrib><creatorcontrib>JI CHENG</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>YAO HONGYAN</au><au>SU FANGXING</au><au>JI CHENG</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Speech synthesis method, device, equipment and medium</title><date>2023-08-29</date><risdate>2023</risdate><abstract>The invention discloses a speech synthesis method and device, equipment and a medium, and relates to the field of big data or finance. The method comprises the following steps: acquiring a voice data set; preprocessing the voice data set to obtain a training data set; training a speech synthesis model according to the training data set; obtaining synthetic data; and inputting the synthesis data into the speech synthesis model to perform speech synthesis. Therefore, the wave network model or the Tarkateon network model is trained through the training data set, and the speech synthesis model can be generated, so that the speech generation model can synthesize speech with high speech naturalness and emotional degree, and the speech synthesis experience of a user is improved. 本申请公开了一种语音合成方法、装置、设备及介质,涉及大数据领域或金融领域。该方法包括:获取语音数据集;对语音数据集进行预处理,以获取训练数据集;根据训练数据集,训练语音合成模型;获取合成数据;将合成数据输入至语音合成模型中,以执行语音合成。由此,通过训练数据集训练波网网络模型或塔科特龙网络模型,可以生成语音合成模型,使得该语音生成模型能够合成语音自然度、情感度高的语音,提升用户的语音合成体验。</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language chi ; eng
recordid cdi_epo_espacenet_CN116665640A
source esp@cenet
subjects ACOUSTICS
CALCULATING
COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
COMPUTING
COUNTING
MUSICAL INSTRUMENTS
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title Speech synthesis method, device, equipment and medium
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-17T11%3A25%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=YAO%20HONGYAN&rft.date=2023-08-29&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN116665640A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true