SPEAKER RECOGNITION DEVICE, SPEAKER RECOGNITION METHOD AND PROGRAM

PROBLEM TO BE SOLVED: To simply and easily perform speaker recognition. SOLUTION: A prediction coefficient for predicting a speech waveform is stored for each speaker in a coefficient storing section 40. A prediction waveform creation section 33 performs arithmetic using a speech data of an input sp...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: HAGA TSUGIHIKO, KONDO TETSUJIRO, TSUJI KOICHIRO, HINO TOSHIYUKI, MACHIMURA MASANORI
Format: Patent
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator HAGA TSUGIHIKO
KONDO TETSUJIRO
TSUJI KOICHIRO
HINO TOSHIYUKI
MACHIMURA MASANORI
description PROBLEM TO BE SOLVED: To simply and easily perform speaker recognition. SOLUTION: A prediction coefficient for predicting a speech waveform is stored for each speaker in a coefficient storing section 40. A prediction waveform creation section 33 performs arithmetic using a speech data of an input speech and the prediction coefficient, and generates a prediction waveform for each speaker. A prediction error calculating section 34 calculates an error of the prediction waveform to the speech waveform of the input speech for each speaker. A speaker specifying section 35 specifies the speaker of the input speech based on the calculated error for each speaker. As for a prediction coefficient, a linear prediction coefficient which is obtained by performing linear prediction analysis on the speech data of the speaker is used. Processing such as converting a time domain to a frequency domain is not needed to perform, and speaker recognition is performed in the time domain. COPYRIGHT: (C)2010,JPO&INPIT
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_JP2009294307A</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>JP2009294307A</sourcerecordid><originalsourceid>FETCH-epo_espacenet_JP2009294307A3</originalsourceid><addsrcrecordid>eNrjZHAKDnB19HYNUghydfZ39_MM8fT3U3BxDfN0dtVRwCbn6xri4e-i4OjnohAQ5O8e5OjLw8CalphTnMoLpbkZlNxcQ5w9dFML8uNTiwsSk1PzUkvivQKMDAwsjSxNjA3MHY2JUgQAVRorpQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>SPEAKER RECOGNITION DEVICE, SPEAKER RECOGNITION METHOD AND PROGRAM</title><source>esp@cenet</source><creator>HAGA TSUGIHIKO ; KONDO TETSUJIRO ; TSUJI KOICHIRO ; HINO TOSHIYUKI ; MACHIMURA MASANORI</creator><creatorcontrib>HAGA TSUGIHIKO ; KONDO TETSUJIRO ; TSUJI KOICHIRO ; HINO TOSHIYUKI ; MACHIMURA MASANORI</creatorcontrib><description>PROBLEM TO BE SOLVED: To simply and easily perform speaker recognition. SOLUTION: A prediction coefficient for predicting a speech waveform is stored for each speaker in a coefficient storing section 40. A prediction waveform creation section 33 performs arithmetic using a speech data of an input speech and the prediction coefficient, and generates a prediction waveform for each speaker. A prediction error calculating section 34 calculates an error of the prediction waveform to the speech waveform of the input speech for each speaker. A speaker specifying section 35 specifies the speaker of the input speech based on the calculated error for each speaker. As for a prediction coefficient, a linear prediction coefficient which is obtained by performing linear prediction analysis on the speech data of the speaker is used. Processing such as converting a time domain to a frequency domain is not needed to perform, and speaker recognition is performed in the time domain. COPYRIGHT: (C)2010,JPO&amp;INPIT</description><language>eng</language><subject>ACOUSTICS ; MUSICAL INSTRUMENTS ; PHYSICS ; SPEECH ANALYSIS OR SYNTHESIS ; SPEECH OR AUDIO CODING OR DECODING ; SPEECH OR VOICE PROCESSING ; SPEECH RECOGNITION</subject><creationdate>2009</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20091217&amp;DB=EPODOC&amp;CC=JP&amp;NR=2009294307A$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,780,885,25564,76547</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20091217&amp;DB=EPODOC&amp;CC=JP&amp;NR=2009294307A$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>HAGA TSUGIHIKO</creatorcontrib><creatorcontrib>KONDO TETSUJIRO</creatorcontrib><creatorcontrib>TSUJI KOICHIRO</creatorcontrib><creatorcontrib>HINO TOSHIYUKI</creatorcontrib><creatorcontrib>MACHIMURA MASANORI</creatorcontrib><title>SPEAKER RECOGNITION DEVICE, SPEAKER RECOGNITION METHOD AND PROGRAM</title><description>PROBLEM TO BE SOLVED: To simply and easily perform speaker recognition. SOLUTION: A prediction coefficient for predicting a speech waveform is stored for each speaker in a coefficient storing section 40. A prediction waveform creation section 33 performs arithmetic using a speech data of an input speech and the prediction coefficient, and generates a prediction waveform for each speaker. A prediction error calculating section 34 calculates an error of the prediction waveform to the speech waveform of the input speech for each speaker. A speaker specifying section 35 specifies the speaker of the input speech based on the calculated error for each speaker. As for a prediction coefficient, a linear prediction coefficient which is obtained by performing linear prediction analysis on the speech data of the speaker is used. Processing such as converting a time domain to a frequency domain is not needed to perform, and speaker recognition is performed in the time domain. COPYRIGHT: (C)2010,JPO&amp;INPIT</description><subject>ACOUSTICS</subject><subject>MUSICAL INSTRUMENTS</subject><subject>PHYSICS</subject><subject>SPEECH ANALYSIS OR SYNTHESIS</subject><subject>SPEECH OR AUDIO CODING OR DECODING</subject><subject>SPEECH OR VOICE PROCESSING</subject><subject>SPEECH RECOGNITION</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2009</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNrjZHAKDnB19HYNUghydfZ39_MM8fT3U3BxDfN0dtVRwCbn6xri4e-i4OjnohAQ5O8e5OjLw8CalphTnMoLpbkZlNxcQ5w9dFML8uNTiwsSk1PzUkvivQKMDAwsjSxNjA3MHY2JUgQAVRorpQ</recordid><startdate>20091217</startdate><enddate>20091217</enddate><creator>HAGA TSUGIHIKO</creator><creator>KONDO TETSUJIRO</creator><creator>TSUJI KOICHIRO</creator><creator>HINO TOSHIYUKI</creator><creator>MACHIMURA MASANORI</creator><scope>EVB</scope></search><sort><creationdate>20091217</creationdate><title>SPEAKER RECOGNITION DEVICE, SPEAKER RECOGNITION METHOD AND PROGRAM</title><author>HAGA TSUGIHIKO ; KONDO TETSUJIRO ; TSUJI KOICHIRO ; HINO TOSHIYUKI ; MACHIMURA MASANORI</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_JP2009294307A3</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng</language><creationdate>2009</creationdate><topic>ACOUSTICS</topic><topic>MUSICAL INSTRUMENTS</topic><topic>PHYSICS</topic><topic>SPEECH ANALYSIS OR SYNTHESIS</topic><topic>SPEECH OR AUDIO CODING OR DECODING</topic><topic>SPEECH OR VOICE PROCESSING</topic><topic>SPEECH RECOGNITION</topic><toplevel>online_resources</toplevel><creatorcontrib>HAGA TSUGIHIKO</creatorcontrib><creatorcontrib>KONDO TETSUJIRO</creatorcontrib><creatorcontrib>TSUJI KOICHIRO</creatorcontrib><creatorcontrib>HINO TOSHIYUKI</creatorcontrib><creatorcontrib>MACHIMURA MASANORI</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>HAGA TSUGIHIKO</au><au>KONDO TETSUJIRO</au><au>TSUJI KOICHIRO</au><au>HINO TOSHIYUKI</au><au>MACHIMURA MASANORI</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>SPEAKER RECOGNITION DEVICE, SPEAKER RECOGNITION METHOD AND PROGRAM</title><date>2009-12-17</date><risdate>2009</risdate><abstract>PROBLEM TO BE SOLVED: To simply and easily perform speaker recognition. SOLUTION: A prediction coefficient for predicting a speech waveform is stored for each speaker in a coefficient storing section 40. A prediction waveform creation section 33 performs arithmetic using a speech data of an input speech and the prediction coefficient, and generates a prediction waveform for each speaker. A prediction error calculating section 34 calculates an error of the prediction waveform to the speech waveform of the input speech for each speaker. A speaker specifying section 35 specifies the speaker of the input speech based on the calculated error for each speaker. As for a prediction coefficient, a linear prediction coefficient which is obtained by performing linear prediction analysis on the speech data of the speaker is used. Processing such as converting a time domain to a frequency domain is not needed to perform, and speaker recognition is performed in the time domain. COPYRIGHT: (C)2010,JPO&amp;INPIT</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng
recordid cdi_epo_espacenet_JP2009294307A
source esp@cenet
subjects ACOUSTICS
MUSICAL INSTRUMENTS
PHYSICS
SPEECH ANALYSIS OR SYNTHESIS
SPEECH OR AUDIO CODING OR DECODING
SPEECH OR VOICE PROCESSING
SPEECH RECOGNITION
title SPEAKER RECOGNITION DEVICE, SPEAKER RECOGNITION METHOD AND PROGRAM
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T17%3A45%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=HAGA%20TSUGIHIKO&rft.date=2009-12-17&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EJP2009294307A%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true