SPEECH RECOGNITION METHOD AND APPARATUS
The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model fo...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Patent |
Sprache: | eng ; fre ; ger |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | WEI, Wei ZHAO, Yinlou ZHANG, Liao JIANG, Zhengxiang FU, Xiaoyin |
description | The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy. |
format | Patent |
fullrecord | <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_EP4075424A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>EP4075424A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_EP4075424A13</originalsourceid><addsrcrecordid>eNrjZFAPDnB1dfZQCHJ19nf38wzx9PdT8HUN8fB3UXD0A-KAAMcgx5DQYB4G1rTEnOJUXijNzaDg5hri7KGbWpAfn1pckJicmpdaEu8aYGJgbmpiZOJoaEyEEgCfkiNo</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><source>esp@cenet</source><creator>WEI, Wei ; ZHAO, Yinlou ; ZHANG, Liao ; JIANG, Zhengxiang ; FU, Xiaoyin</creator><creatorcontrib>WEI, Wei ; ZHAO, Yinlou ; ZHANG, Liao ; JIANG, Zhengxiang ; FU, Xiaoyin</creatorcontrib><description>The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.</description><language>eng ; fre ; ger</language><subject>ACOUSTICS ; GYROSCOPIC INSTRUMENTS ; MEASURING ; MEASURING DISTANCES, LEVELS OR BEARINGS ; MUSICAL INSTRUMENTS ; NAVIGATION ; PHOTOGRAMMETRY OR VIDEOGRAMMETRY ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION ; SURVEYING ; TESTING</subject><creationdate>2022</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20221019&DB=EPODOC&CC=EP&NR=4075424A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20221019&DB=EPODOC&CC=EP&NR=4075424A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>WEI, Wei</creatorcontrib><creatorcontrib>ZHAO, Yinlou</creatorcontrib><creatorcontrib>ZHANG, Liao</creatorcontrib><creatorcontrib>JIANG, Zhengxiang</creatorcontrib><creatorcontrib>FU, Xiaoyin</creatorcontrib><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><description>The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.</description><subject>ACOUSTICS</subject><subject>GYROSCOPIC INSTRUMENTS</subject><subject>MEASURING</subject><subject>MEASURING DISTANCES, LEVELS OR BEARINGS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>NAVIGATION</subject><subject>PHOTOGRAMMETRY OR VIDEOGRAMMETRY</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><subject>SURVEYING</subject><subject>TESTING</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2022</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZFAPDnB1dfZQCHJ19nf38wzx9PdT8HUN8fB3UXD0A-KAAMcgx5DQYB4G1rTEnOJUXijNzaDg5hri7KGbWpAfn1pckJicmpdaEu8aYGJgbmpiZOJoaEyEEgCfkiNo</recordid><startdate>20221019</startdate><enddate>20221019</enddate><creator>WEI, Wei</creator><creator>ZHAO, Yinlou</creator><creator>ZHANG, Liao</creator><creator>JIANG, Zhengxiang</creator><creator>FU, Xiaoyin</creator><scope>EVB</scope></search><sort><creationdate>20221019</creationdate><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><author>WEI, Wei ; ZHAO, Yinlou ; ZHANG, Liao ; JIANG, Zhengxiang ; FU, Xiaoyin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_EP4075424A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng ; fre ; ger</language><creationdate>2022</creationdate><topic>ACOUSTICS</topic><topic>GYROSCOPIC INSTRUMENTS</topic><topic>MEASURING</topic><topic>MEASURING DISTANCES, LEVELS OR BEARINGS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>NAVIGATION</topic><topic>PHOTOGRAMMETRY OR VIDEOGRAMMETRY</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><topic>SURVEYING</topic><topic>TESTING</topic><toplevel>online_resources</toplevel><creatorcontrib>WEI, Wei</creatorcontrib><creatorcontrib>ZHAO, Yinlou</creatorcontrib><creatorcontrib>ZHANG, Liao</creatorcontrib><creatorcontrib>JIANG, Zhengxiang</creatorcontrib><creatorcontrib>FU, Xiaoyin</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>WEI, Wei</au><au>ZHAO, Yinlou</au><au>ZHANG, Liao</au><au>JIANG, Zhengxiang</au><au>FU, Xiaoyin</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>SPEECH RECOGNITION METHOD AND APPARATUS</title><date>2022-10-19</date><risdate>2022</risdate><abstract>The present disclosure discloses a speech recognition method and apparatus, and relates to the field of speech and deep learning technologies. A specific implementation scheme involves: acquiring candidate recognition results with first N recognition scores outputted by a speech recognition model for to-be-recognized speech, N being a positive integer greater than 1; scoring the N candidate recognition results based on pronunciation similarities between candidate recognition results and pre-collected popular entities, to obtain similarity scores of the candidate recognition results; and integrating the recognition scores and the similarity scores of the candidate recognition results to determine a recognition result corresponding to the to-be-recognized speech from the N candidate recognition results. The present disclosure can improve recognition accuracy.</abstract><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | |
ispartof | |
issn | |
language | eng ; fre ; ger |
recordid | cdi_epo_espacenet_EP4075424A1 |
source | esp@cenet |
subjects | ACOUSTICS GYROSCOPIC INSTRUMENTS MEASURING MEASURING DISTANCES, LEVELS OR BEARINGS MUSICAL INSTRUMENTS NAVIGATION PHOTOGRAMMETRY OR VIDEOGRAMMETRY PHYSICS SPEECH ANALYSIS OR SYNTHESIS SPEECH OR AUDIO CODING OR DECODING SPEECH OR VOICE PROCESSING SPEECH RECOGNITION SURVEYING TESTING |
title | SPEECH RECOGNITION METHOD AND APPARATUS |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-24T22%3A18%3A12IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=WEI,%20Wei&rft.date=2022-10-19&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EEP4075424A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |