SPEECH RECOGNITION METHOD AND APPARATUS

The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model fo...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: WEI, Wei, ZHAO, Yinlou, ZHANG, Liao, JIANG, Zhengxiang, FU, Xiaoyin
Format: Patent
Sprache:eng ; fre ; ger
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator WEI, Wei
ZHAO, Yinlou
ZHANG, Liao
JIANG, Zhengxiang
FU, Xiaoyin
description The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_EP4075424A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>EP4075424A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_EP4075424A13</originalsourceid><addsrcrecordid>eNrjZFAPDnB1dfZQCHJ19nf38wzx9PdT8HUN8fB3UXD0A-KAAMcgx5DQYB4G1rTEnOJUXijNzaDg5hri7KGbWpAfn1pckJicmpdaEu8aYGJgbmpiZOJoaEyEEgCfkiNo</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><source>esp@cenet</source><creator>WEI, Wei ; ZHAO, Yinlou ; ZHANG, Liao ; JIANG, Zhengxiang ; FU, Xiaoyin</creator><creatorcontrib>WEI, Wei ; ZHAO, Yinlou ; ZHANG, Liao ; JIANG, Zhengxiang ; FU, Xiaoyin</creatorcontrib><description>The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.</description><language>eng ; fre ; ger</language><subject>ACOUSTICS ; GYROSCOPIC INSTRUMENTS ; MEASURING ; MEASURING DISTANCES, LEVELS OR BEARINGS ; MUSICAL INSTRUMENTS ; NAVIGATION ; PHOTOGRAMMETRY OR VIDEOGRAMMETRY ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION ; SURVEYING ; TESTING</subject><creationdate>2022</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20221019&amp;DB=EPODOC&amp;CC=EP&amp;NR=4075424A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20221019&amp;DB=EPODOC&amp;CC=EP&amp;NR=4075424A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>WEI, Wei</creatorcontrib><creatorcontrib>ZHAO, Yinlou</creatorcontrib><creatorcontrib>ZHANG, Liao</creatorcontrib><creatorcontrib>JIANG, Zhengxiang</creatorcontrib><creatorcontrib>FU, Xiaoyin</creatorcontrib><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><description>The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.</description><subject>ACOUSTICS</subject><subject>GYROSCOPIC INSTRUMENTS</subject><subject>MEASURING</subject><subject>MEASURING DISTANCES, LEVELS OR BEARINGS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>NAVIGATION</subject><subject>PHOTOGRAMMETRY OR VIDEOGRAMMETRY</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><subject>SURVEYING</subject><subject>TESTING</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2022</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZFAPDnB1dfZQCHJ19nf38wzx9PdT8HUN8fB3UXD0A-KAAMcgx5DQYB4G1rTEnOJUXijNzaDg5hri7KGbWpAfn1pckJicmpdaEu8aYGJgbmpiZOJoaEyEEgCfkiNo</recordid><startdate>20221019</startdate><enddate>20221019</enddate><creator>WEI, Wei</creator><creator>ZHAO, Yinlou</creator><creator>ZHANG, Liao</creator><creator>JIANG, Zhengxiang</creator><creator>FU, Xiaoyin</creator><scope>EVB</scope></search><sort><creationdate>20221019</creationdate><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><author>WEI, Wei ; ZHAO, Yinlou ; ZHANG, Liao ; JIANG, Zhengxiang ; FU, Xiaoyin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_EP4075424A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng ; fre ; ger</language><creationdate>2022</creationdate><topic>ACOUSTICS</topic><topic>GYROSCOPIC INSTRUMENTS</topic><topic>MEASURING</topic><topic>MEASURING DISTANCES, LEVELS OR BEARINGS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>NAVIGATION</topic><topic>PHOTOGRAMMETRY OR VIDEOGRAMMETRY</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><topic>SURVEYING</topic><topic>TESTING</topic><toplevel>online_resources</toplevel><creatorcontrib>WEI, Wei</creatorcontrib><creatorcontrib>ZHAO, Yinlou</creatorcontrib><creatorcontrib>ZHANG, Liao</creatorcontrib><creatorcontrib>JIANG, Zhengxiang</creatorcontrib><creatorcontrib>FU, Xiaoyin</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>WEI, Wei</au><au>ZHAO, Yinlou</au><au>ZHANG, Liao</au><au>JIANG, Zhengxiang</au><au>FU, Xiaoyin</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><date>2022-10-19</date><risdate>2022</risdate><abstract>The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng ; fre ; ger
recordid cdi_epo_espacenet_EP4075424A1
source esp@cenet
subjects ACOUSTICS
GYROSCOPIC INSTRUMENTS
MEASURING
MEASURING DISTANCES, LEVELS OR BEARINGS
MUSICAL INSTRUMENTS
NAVIGATION
PHOTOGRAMMETRY OR VIDEOGRAMMETRY
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
SURVEYING
TESTING
title SPEECH RECOGNITION METHOD AND APPARATUS
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-24T22%3A18%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=WEI,%20Wei&rft.date=2022-10-19&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EEP4075424A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true