DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING

Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Traynor, Patrick G, Butler, Kevin, Gates, Carrie, Warren, Kevin S, Olszewski, Daniel, Layton, Seth
Format: Patent
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Traynor, Patrick G
Butler, Kevin
Gates, Carrie
Warren, Kevin S
Olszewski, Daniel
Layton, Seth
description Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_US2023343342A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>US2023343342A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_US2023343342A13</originalsourceid><addsrcrecordid>eNrjZLBxcQ1xdQ7x9HNXcAx18fRXcHF1DXBz9HYNVgjxCPIPdfdQcHT2Dw0O8XRWCAjyD_Z3ATJ8_V1cfYBaeBhY0xJzilN5oTQ3g7Kba4izh25qQX58anFBYnJqXmpJfGiwkYGRsbEJEBk5GhoTpwoAsykqig</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><source>esp@cenet</source><creator>Traynor, Patrick G ; Butler, Kevin ; Gates, Carrie ; Warren, Kevin S ; Olszewski, Daniel ; Layton, Seth</creator><creatorcontrib>Traynor, Patrick G ; Butler, Kevin ; Gates, Carrie ; Warren, Kevin S ; Olszewski, Daniel ; Layton, Seth</creatorcontrib><description>Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.</description><language>eng</language><subject>ACOUSTICS ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2023</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20231026&amp;DB=EPODOC&amp;CC=US&amp;NR=2023343342A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76289</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20231026&amp;DB=EPODOC&amp;CC=US&amp;NR=2023343342A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>Traynor, Patrick G</creatorcontrib><creatorcontrib>Butler, Kevin</creatorcontrib><creatorcontrib>Gates, Carrie</creatorcontrib><creatorcontrib>Warren, Kevin S</creatorcontrib><creatorcontrib>Olszewski, Daniel</creatorcontrib><creatorcontrib>Layton, Seth</creatorcontrib><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><description>Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.</description><subject>ACOUSTICS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2023</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZLBxcQ1xdQ7x9HNXcAx18fRXcHF1DXBz9HYNVgjxCPIPdfdQcHT2Dw0O8XRWCAjyD_Z3ATJ8_V1cfYBaeBhY0xJzilN5oTQ3g7Kba4izh25qQX58anFBYnJqXmpJfGiwkYGRsbEJEBk5GhoTpwoAsykqig</recordid><startdate>20231026</startdate><enddate>20231026</enddate><creator>Traynor, Patrick G</creator><creator>Butler, Kevin</creator><creator>Gates, Carrie</creator><creator>Warren, Kevin S</creator><creator>Olszewski, Daniel</creator><creator>Layton, Seth</creator><scope>EVB</scope></search><sort><creationdate>20231026</creationdate><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><author>Traynor, Patrick G ; Butler, Kevin ; Gates, Carrie ; Warren, Kevin S ; Olszewski, Daniel ; Layton, Seth</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_US2023343342A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng</language><creationdate>2023</creationdate><topic>ACOUSTICS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>Traynor, Patrick G</creatorcontrib><creatorcontrib>Butler, Kevin</creatorcontrib><creatorcontrib>Gates, Carrie</creatorcontrib><creatorcontrib>Warren, Kevin S</creatorcontrib><creatorcontrib>Olszewski, Daniel</creatorcontrib><creatorcontrib>Layton, Seth</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Traynor, Patrick G</au><au>Butler, Kevin</au><au>Gates, Carrie</au><au>Warren, Kevin S</au><au>Olszewski, Daniel</au><au>Layton, Seth</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><date>2023-10-26</date><risdate>2023</risdate><abstract>Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng
recordid cdi_epo_espacenet_US2023343342A1
source esp@cenet
subjects ACOUSTICS
MUSICAL INSTRUMENTS
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T16%3A52%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=Traynor,%20Patrick%20G&rft.date=2023-10-26&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EUS2023343342A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true