DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING
Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic a...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Patent |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Traynor, Patrick G Butler, Kevin Gates, Carrie Warren, Kevin S Olszewski, Daniel Layton, Seth |
description | Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes. |
format | Patent |
fullrecord | <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_US2023343342A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>US2023343342A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_US2023343342A13</originalsourceid><addsrcrecordid>eNrjZLBxcQ1xdQ7x9HNXcAx18fRXcHF1DXBz9HYNVgjxCPIPdfdQcHT2Dw0O8XRWCAjyD_Z3ATJ8_V1cfYBaeBhY0xJzilN5oTQ3g7Kba4izh25qQX58anFBYnJqXmpJfGiwkYGRsbEJEBk5GhoTpwoAsykqig</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><source>esp@cenet</source><creator>Traynor, Patrick G ; Butler, Kevin ; Gates, Carrie ; Warren, Kevin S ; Olszewski, Daniel ; Layton, Seth</creator><creatorcontrib>Traynor, Patrick G ; Butler, Kevin ; Gates, Carrie ; Warren, Kevin S ; Olszewski, Daniel ; Layton, Seth</creatorcontrib><description>Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.</description><language>eng</language><subject>ACOUSTICS ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2023</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20231026&DB=EPODOC&CC=US&NR=2023343342A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76289</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&date=20231026&DB=EPODOC&CC=US&NR=2023343342A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>Traynor, Patrick G</creatorcontrib><creatorcontrib>Butler, Kevin</creatorcontrib><creatorcontrib>Gates, Carrie</creatorcontrib><creatorcontrib>Warren, Kevin S</creatorcontrib><creatorcontrib>Olszewski, Daniel</creatorcontrib><creatorcontrib>Layton, Seth</creatorcontrib><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><description>Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.</description><subject>ACOUSTICS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2023</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZLBxcQ1xdQ7x9HNXcAx18fRXcHF1DXBz9HYNVgjxCPIPdfdQcHT2Dw0O8XRWCAjyD_Z3ATJ8_V1cfYBaeBhY0xJzilN5oTQ3g7Kba4izh25qQX58anFBYnJqXmpJfGiwkYGRsbEJEBk5GhoTpwoAsykqig</recordid><startdate>20231026</startdate><enddate>20231026</enddate><creator>Traynor, Patrick G</creator><creator>Butler, Kevin</creator><creator>Gates, Carrie</creator><creator>Warren, Kevin S</creator><creator>Olszewski, Daniel</creator><creator>Layton, Seth</creator><scope>EVB</scope></search><sort><creationdate>20231026</creationdate><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><author>Traynor, Patrick G ; Butler, Kevin ; Gates, Carrie ; Warren, Kevin S ; Olszewski, Daniel ; Layton, Seth</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_US2023343342A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng</language><creationdate>2023</creationdate><topic>ACOUSTICS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>Traynor, Patrick G</creatorcontrib><creatorcontrib>Butler, Kevin</creatorcontrib><creatorcontrib>Gates, Carrie</creatorcontrib><creatorcontrib>Warren, Kevin S</creatorcontrib><creatorcontrib>Olszewski, Daniel</creatorcontrib><creatorcontrib>Layton, Seth</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Traynor, Patrick G</au><au>Butler, Kevin</au><au>Gates, Carrie</au><au>Warren, Kevin S</au><au>Olszewski, Daniel</au><au>Layton, Seth</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING</title><date>2023-10-26</date><risdate>2023</risdate><abstract>Various embodiments of the present disclosure provide for detecting audio deepfakes through acoustic prosodic modeling. In one example, an embodiment provides for extracting one or more prosodic features from an audio sample and classifying the audio sample as a deepfake audio sample or an organic audio sample by applying a machine learning model to the one or more prosodic features. The one or more prosodic features can be indicative of one or more prosodic characteristics associated with human speech. Additionally, the machine learning model can be configured as a classification-based detector for audio deepfakes.</abstract><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | |
ispartof | |
issn | |
language | eng |
recordid | cdi_epo_espacenet_US2023343342A1 |
source | esp@cenet |
subjects | ACOUSTICS MUSICAL INSTRUMENTS PHYSICS SPEECH ANALYSIS OR SYNTHESIS SPEECH OR AUDIO CODING OR DECODING SPEECH OR VOICE PROCESSING SPEECH RECOGNITION |
title | DETECTING AUDIO DEEPFAKES THROUGH ACOUSTIC PROSODIC MODELING |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T16%3A52%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=Traynor,%20Patrick%20G&rft.date=2023-10-26&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EUS2023343342A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |