Vowel classification with combining pitch detection and one‐dimensional convolutional neural network based classifier for gender identification

Human speech signals may contain specific information regarding a speaker's characteristics, and these signals can be very useful in applications involving interactive voice response (IVR) and automatic speech recognition (ASR). For IVR and ASR applications, speaker classification into differen...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IET signal processing 2023-05, Vol.17 (5), p.n/a
Hauptverfasser: Lin, Chia‐Hung, Lai, Hsiang‐Yueh, Huang, Ping‐Tzan, Chen, Pi‐Yun, Li, Chien‐Ming
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Human speech signals may contain specific information regarding a speaker's characteristics, and these signals can be very useful in applications involving interactive voice response (IVR) and automatic speech recognition (ASR). For IVR and ASR applications, speaker classification into different ages and gender groups can be applied in human–machine interaction or computer‐based interaction systems for customised advertisement, translation (text generation), machine dialog systems, or self‐service applications. Hence, an IVR‐based system dictates that ASR should function through users' voices (specific voice‐frequency bands) to identify customers' age and gender and interact with a host system. In the present study, we intended to combine a pitch detection (PD)‐based extractor and a voice classifier for gender identification. The Yet Another Algorithm for Pitch Tracking (YAAPT)‐based PD method was designed to extract the voice fundamental frequency (F0) from non‐stationary speaker's voice signals, allowing us to achieve gender identification, by distinguishing differences in F0 between adult females and males, and classify voices into adult and children groups. Then, in vowel voice signal classification, a one‐dimensional (1D) convolutional neural network (CNN), consisted of a multi‐round 1D kernel convolutional layer, a 1D pooling process, and a vowel classifier that could preliminary divide feature patterns into three level ranges of F0, including adult and children groups. Consequently, a classifier was used in the classification layer to identify the speakers' gender. The proposed PD‐based extractor and voice classifier could reduce complexity and improve classification efficiency. Acoustic datasets were selected from the Hillenbrand database for experimental tests on 12 vowels classifications, and K‐fold cross‐validations were performed. The experimental results demonstrated that our approach is a very promising method to quantify the proposed classifier's performance in terms of recall (%), precision (%), accuracy (%), and F1 score. The YAAPT‐based pitch detection algorithm is used to design an extractor to extract the spectral peak (F0) patterns, and 1D CNN‐based classifier is employed to use the feature pattern to identify females or males in adult or children groups.
ISSN:1751-9675
1751-9683
DOI:10.1049/sil2.12216