VOICE TO TEXT CONVERSION BASED ON THIRD-PARTY AGENT CONTENT

Implementations relate to dynamically, and in a context-sensitive manner, biasing voice to text conversion. In some implementations, the biasing of voice to text conversions is performed by a voice to text engine of a local agent, and the biasing is based at least in part on content provided to the...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: VEMURI, Sunil, DOGRULTAN, Ertan, KIRAZCI, Ulas, WANG, Bo, ALEKSIC, Petar, JAMES, Barnaby, SCHAIRER, David
Format: Patent
Sprache:eng ; fre ; ger
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator VEMURI, Sunil
DOGRULTAN, Ertan
KIRAZCI, Ulas
WANG, Bo
ALEKSIC, Petar
JAMES, Barnaby
SCHAIRER, David
description Implementations relate to dynamically, and in a context-sensitive manner, biasing voice to text conversion. In some implementations, the biasing of voice to text conversions is performed by a voice to text engine of a local agent, and the biasing is based at least in part on content provided to the local agent by a third-party (3P) agent that is in network communication with the local agent. In some of those implementations, the content includes contextual parameters that are provided by the 3P agent in combination with responsive content generated by the 3P agent during a dialog that: is between the 3P agent, and a user of a voice-enabled electronic device; and is facilitated by the local agent. The contextual parameters indicate potential feature(s) of further voice input that is to be provided in response to the responsive content generated by the 3P agent. Implementations described herein reduce the use of various computational resources that may otherwise be consumed by inaccurate representations of voice inputs (e.g., network traffic consumed by additional "turns" that may be necessary to correct inaccurate representations of voice input).
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_EP4332959A2</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>EP4332959A2</sourcerecordid><originalsourceid>FETCH-epo_espacenet_EP4332959A23</originalsourceid><addsrcrecordid>eNrjZLAO8_d0dlUI8VcIcY0IUXD29wtzDQr29PdTcHIMdnVRADJCPDyDXHQDHINCIhUc3V39wKpCgDQPA2taYk5xKi-U5mZQcHMNcfbQTS3Ij08tLkhMTs1LLYl3DTAxNjayNLV0NDImQgkAjHYo9A</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>VOICE TO TEXT CONVERSION BASED ON THIRD-PARTY AGENT CONTENT</title><source>esp@cenet</source><creator>VEMURI, Sunil ; DOGRULTAN, Ertan ; KIRAZCI, Ulas ; WANG, Bo ; ALEKSIC, Petar ; JAMES, Barnaby ; SCHAIRER, David</creator><creatorcontrib>VEMURI, Sunil ; DOGRULTAN, Ertan ; KIRAZCI, Ulas ; WANG, Bo ; ALEKSIC, Petar ; JAMES, Barnaby ; SCHAIRER, David</creatorcontrib><description>Implementations relate to dynamically, and in a context-sensitive manner, biasing voice to text conversion. In some implementations, the biasing of voice to text conversions is performed by a voice to text engine of a local agent, and the biasing is based at least in part on content provided to the local agent by a third-party (3P) agent that is in network communication with the local agent. In some of those implementations, the content includes contextual parameters that are provided by the 3P agent in combination with responsive content generated by the 3P agent during a dialog that: is between the 3P agent, and a user of a voice-enabled electronic device; and is facilitated by the local agent. The contextual parameters indicate potential feature(s) of further voice input that is to be provided in response to the responsive content generated by the 3P agent. Implementations described herein reduce the use of various computational resources that may otherwise be consumed by inaccurate representations of voice inputs (e.g., network traffic consumed by additional "turns" that may be necessary to correct inaccurate representations of voice input).</description><language>eng ; fre ; ger</language><subject>ACOUSTICS ; CALCULATING ; COMPUTING ; COUNTING ; ELECTRIC DIGITAL DATA PROCESSING ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240306&amp;DB=EPODOC&amp;CC=EP&amp;NR=4332959A2$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240306&amp;DB=EPODOC&amp;CC=EP&amp;NR=4332959A2$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>VEMURI, Sunil</creatorcontrib><creatorcontrib>DOGRULTAN, Ertan</creatorcontrib><creatorcontrib>KIRAZCI, Ulas</creatorcontrib><creatorcontrib>WANG, Bo</creatorcontrib><creatorcontrib>ALEKSIC, Petar</creatorcontrib><creatorcontrib>JAMES, Barnaby</creatorcontrib><creatorcontrib>SCHAIRER, David</creatorcontrib><title>VOICE TO TEXT CONVERSION BASED ON THIRD-PARTY AGENT CONTENT</title><description>Implementations relate to dynamically, and in a context-sensitive manner, biasing voice to text conversion. In some implementations, the biasing of voice to text conversions is performed by a voice to text engine of a local agent, and the biasing is based at least in part on content provided to the local agent by a third-party (3P) agent that is in network communication with the local agent. In some of those implementations, the content includes contextual parameters that are provided by the 3P agent in combination with responsive content generated by the 3P agent during a dialog that: is between the 3P agent, and a user of a voice-enabled electronic device; and is facilitated by the local agent. The contextual parameters indicate potential feature(s) of further voice input that is to be provided in response to the responsive content generated by the 3P agent. Implementations described herein reduce the use of various computational resources that may otherwise be consumed by inaccurate representations of voice inputs (e.g., network traffic consumed by additional "turns" that may be necessary to correct inaccurate representations of voice input).</description><subject>ACOUSTICS</subject><subject>CALCULATING</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>ELECTRIC DIGITAL DATA PROCESSING</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZLAO8_d0dlUI8VcIcY0IUXD29wtzDQr29PdTcHIMdnVRADJCPDyDXHQDHINCIhUc3V39wKpCgDQPA2taYk5xKi-U5mZQcHMNcfbQTS3Ij08tLkhMTs1LLYl3DTAxNjayNLV0NDImQgkAjHYo9A</recordid><startdate>20240306</startdate><enddate>20240306</enddate><creator>VEMURI, Sunil</creator><creator>DOGRULTAN, Ertan</creator><creator>KIRAZCI, Ulas</creator><creator>WANG, Bo</creator><creator>ALEKSIC, Petar</creator><creator>JAMES, Barnaby</creator><creator>SCHAIRER, David</creator><scope>EVB</scope></search><sort><creationdate>20240306</creationdate><title>VOICE TO TEXT CONVERSION BASED ON THIRD-PARTY AGENT CONTENT</title><author>VEMURI, Sunil ; DOGRULTAN, Ertan ; KIRAZCI, Ulas ; WANG, Bo ; ALEKSIC, Petar ; JAMES, Barnaby ; SCHAIRER, David</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_EP4332959A23</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng ; fre ; ger</language><creationdate>2024</creationdate><topic>ACOUSTICS</topic><topic>CALCULATING</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>ELECTRIC DIGITAL DATA PROCESSING</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>VEMURI, Sunil</creatorcontrib><creatorcontrib>DOGRULTAN, Ertan</creatorcontrib><creatorcontrib>KIRAZCI, Ulas</creatorcontrib><creatorcontrib>WANG, Bo</creatorcontrib><creatorcontrib>ALEKSIC, Petar</creatorcontrib><creatorcontrib>JAMES, Barnaby</creatorcontrib><creatorcontrib>SCHAIRER, David</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>VEMURI, Sunil</au><au>DOGRULTAN, Ertan</au><au>KIRAZCI, Ulas</au><au>WANG, Bo</au><au>ALEKSIC, Petar</au><au>JAMES, Barnaby</au><au>SCHAIRER, David</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>VOICE TO TEXT CONVERSION BASED ON THIRD-PARTY AGENT CONTENT</title><date>2024-03-06</date><risdate>2024</risdate><abstract>Implementations relate to dynamically, and in a context-sensitive manner, biasing voice to text conversion. In some implementations, the biasing of voice to text conversions is performed by a voice to text engine of a local agent, and the biasing is based at least in part on content provided to the local agent by a third-party (3P) agent that is in network communication with the local agent. In some of those implementations, the content includes contextual parameters that are provided by the 3P agent in combination with responsive content generated by the 3P agent during a dialog that: is between the 3P agent, and a user of a voice-enabled electronic device; and is facilitated by the local agent. The contextual parameters indicate potential feature(s) of further voice input that is to be provided in response to the responsive content generated by the 3P agent. Implementations described herein reduce the use of various computational resources that may otherwise be consumed by inaccurate representations of voice inputs (e.g., network traffic consumed by additional "turns" that may be necessary to correct inaccurate representations of voice input).</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng ; fre ; ger
recordid cdi_epo_espacenet_EP4332959A2
source esp@cenet
subjects ACOUSTICS
CALCULATING
COMPUTING
COUNTING
ELECTRIC DIGITAL DATA PROCESSING
MUSICAL INSTRUMENTS
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title VOICE TO TEXT CONVERSION BASED ON THIRD-PARTY AGENT CONTENT
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T13%3A18%3A29IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=VEMURI,%20Sunil&rft.date=2024-03-06&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EEP4332959A2%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true