Voice-to-text conversion based on third party proxy content
The invention discloses voice-to-text conversion based on third-party proxy content, and relates to dynamic and context-sensitive biasing of voice-to-text conversion. In some implementations, the biasing of voice-to-text conversion is performed by a voice-to-text engine of a home agent, and the bias...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Patent |
Sprache: | chi ; eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | KIRAZCI ULAS VEMURI, SUNIL JAMES BARNABY ALEKSIC, PETRA DOROURTAN, ERTAN WANG BO SCHERER, DAVID |
description | The invention discloses voice-to-text conversion based on third-party proxy content, and relates to dynamic and context-sensitive biasing of voice-to-text conversion. In some implementations, the biasing of voice-to-text conversion is performed by a voice-to-text engine of a home agent, and the biasing is based at least in part on content provided to the home agent by a third party (3P) agent in network communication with the home agent. In some implementations, the content includes context parameters provided by the 3P agent in combination with response content generated by the 3P agent during a conversation between the 3P agent and a user of the voice-enabled electronic device; and assistance is realized by the local agent. The contextual parameters indicate potential characteristics of additional voice input to be provided in response to response content generated by the 3P agent. Embodiments reduce the use of various computing resources (e.g., network traffic that may be consumed by additional rounds nece |
format | Patent |
fullrecord | <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_CN117059097A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>CN117059097A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_CN117059097A3</originalsourceid><addsrcrecordid>eNrjZLAOy89MTtUtydctSa0oUUjOzytLLSrOzM9TSEosTk1RADJKMjKLUhQKEotKKhUKivIrKkGqSlLzSngYWNMSc4pTeaE0N4Oim2uIs4duakF-fGpxQWJyal5qSbyzn6GhuYGppYGluaMxMWoAxMwvsg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>Voice-to-text conversion based on third party proxy content</title><source>esp@cenet</source><creator>KIRAZCI ULAS ; VEMURI, SUNIL ; JAMES BARNABY ; ALEKSIC, PETRA ; DOROURTAN, ERTAN ; WANG BO ; SCHERER, DAVID</creator><creatorcontrib>KIRAZCI ULAS ; VEMURI, SUNIL ; JAMES BARNABY ; ALEKSIC, PETRA ; DOROURTAN, ERTAN ; WANG BO ; SCHERER, DAVID</creatorcontrib><description>The invention discloses voice-to-text conversion based on third-party proxy content, and relates to dynamic and context-sensitive biasing of voice-to-text conversion. In some implementations, the biasing of voice-to-text conversion is performed by a voice-to-text engine of a home agent, and the biasing is based at least in part on content provided to the home agent by a third party (3P) agent in network communication with the home agent. In some implementations, the content includes context parameters provided by the 3P agent in combination with response content generated by the 3P agent during a conversation between the 3P agent and a user of the voice-enabled electronic device; and assistance is realized by the local agent. The contextual parameters indicate potential characteristics of additional voice input to be provided in response to response content generated by the 3P agent. Embodiments reduce the use of various computing resources (e.g., network traffic that may be consumed by additional rounds nece</description><language>chi ; eng</language><subject>ACOUSTICS ; CALCULATING ; COMPUTING ; COUNTING ; ELECTRIC DIGITAL DATA PROCESSING ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2023</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20231114&DB=EPODOC&CC=CN&NR=117059097A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76289</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20231114&DB=EPODOC&CC=CN&NR=117059097A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>KIRAZCI ULAS</creatorcontrib><creatorcontrib>VEMURI, SUNIL</creatorcontrib><creatorcontrib>JAMES BARNABY</creatorcontrib><creatorcontrib>ALEKSIC, PETRA</creatorcontrib><creatorcontrib>DOROURTAN, ERTAN</creatorcontrib><creatorcontrib>WANG BO</creatorcontrib><creatorcontrib>SCHERER, DAVID</creatorcontrib><title>Voice-to-text conversion based on third party proxy content</title><description>The invention discloses voice-to-text conversion based on third-party proxy content, and relates to dynamic and context-sensitive biasing of voice-to-text conversion. In some implementations, the biasing of voice-to-text conversion is performed by a voice-to-text engine of a home agent, and the biasing is based at least in part on content provided to the home agent by a third party (3P) agent in network communication with the home agent. In some implementations, the content includes context parameters provided by the 3P agent in combination with response content generated by the 3P agent during a conversation between the 3P agent and a user of the voice-enabled electronic device; and assistance is realized by the local agent. The contextual parameters indicate potential characteristics of additional voice input to be provided in response to response content generated by the 3P agent. Embodiments reduce the use of various computing resources (e.g., network traffic that may be consumed by additional rounds nece</description><subject>ACOUSTICS</subject><subject>CALCULATING</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>ELECTRIC DIGITAL DATA PROCESSING</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2023</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZLAOy89MTtUtydctSa0oUUjOzytLLSrOzM9TSEosTk1RADJKMjKLUhQKEotKKhUKivIrKkGqSlLzSngYWNMSc4pTeaE0N4Oim2uIs4duakF-fGpxQWJyal5qSbyzn6GhuYGppYGluaMxMWoAxMwvsg</recordid><startdate>20231114</startdate><enddate>20231114</enddate><creator>KIRAZCI ULAS</creator><creator>VEMURI, SUNIL</creator><creator>JAMES BARNABY</creator><creator>ALEKSIC, PETRA</creator><creator>DOROURTAN, ERTAN</creator><creator>WANG BO</creator><creator>SCHERER, DAVID</creator><scope>EVB</scope></search><sort><creationdate>20231114</creationdate><title>Voice-to-text conversion based on third party proxy content</title><author>KIRAZCI ULAS ; VEMURI, SUNIL ; JAMES BARNABY ; ALEKSIC, PETRA ; DOROURTAN, ERTAN ; WANG BO ; SCHERER, DAVID</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_CN117059097A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>chi ; eng</language><creationdate>2023</creationdate><topic>ACOUSTICS</topic><topic>CALCULATING</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>ELECTRIC DIGITAL DATA PROCESSING</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>KIRAZCI ULAS</creatorcontrib><creatorcontrib>VEMURI, SUNIL</creatorcontrib><creatorcontrib>JAMES BARNABY</creatorcontrib><creatorcontrib>ALEKSIC, PETRA</creatorcontrib><creatorcontrib>DOROURTAN, ERTAN</creatorcontrib><creatorcontrib>WANG BO</creatorcontrib><creatorcontrib>SCHERER, DAVID</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>KIRAZCI ULAS</au><au>VEMURI, SUNIL</au><au>JAMES BARNABY</au><au>ALEKSIC, PETRA</au><au>DOROURTAN, ERTAN</au><au>WANG BO</au><au>SCHERER, DAVID</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>Voice-to-text conversion based on third party proxy content</title><date>2023-11-14</date><risdate>2023</risdate><abstract>The invention discloses voice-to-text conversion based on third-party proxy content, and relates to dynamic and context-sensitive biasing of voice-to-text conversion. In some implementations, the biasing of voice-to-text conversion is performed by a voice-to-text engine of a home agent, and the biasing is based at least in part on content provided to the home agent by a third party (3P) agent in network communication with the home agent. In some implementations, the content includes context parameters provided by the 3P agent in combination with response content generated by the 3P agent during a conversation between the 3P agent and a user of the voice-enabled electronic device; and assistance is realized by the local agent. The contextual parameters indicate potential characteristics of additional voice input to be provided in response to response content generated by the 3P agent. Embodiments reduce the use of various computing resources (e.g., network traffic that may be consumed by additional rounds nece</abstract><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | |
ispartof | |
issn | |
language | chi ; eng |
recordid | cdi_epo_espacenet_CN117059097A |
source | esp@cenet |
subjects | ACOUSTICS CALCULATING COMPUTING COUNTING ELECTRIC DIGITAL DATA PROCESSING MUSICAL INSTRUMENTS PHYSICS SPEECH ANALYSIS OR SYNTHESIS SPEECH OR AUDIO CODING OR DECODING SPEECH OR VOICE PROCESSING SPEECH RECOGNITION |
title | Voice-to-text conversion based on third party proxy content |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-09T11%3A55%3A14IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=KIRAZCI%20ULAS&rft.date=2023-11-14&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3ECN117059097A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |