AI Generated Music Using Speech Emotion Recognition

This study aims to compare two different implementations of speech emotion recognition models.The emphasis is directed towards evaluating their efficacy in capturing and characterizing dialogues portrayed by actors within a film scene to create suitable musical intervals. The goal of the overarching...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Murru, Roberto, Krug, Jonas, Schmid, Tom, Steba, Garri, Giacinto, Giorgio, von Hoffmann, Alexander
Format: Dataset
Sprache:eng
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:This study aims to compare two different implementations of speech emotion recognition models.The emphasis is directed towards evaluating their efficacy in capturing and characterizing dialogues portrayed by actors within a film scene to create suitable musical intervals. The goal of the overarching research intends to derive indications to enhance the compositional process of film scores by recognizing the emotion in a particular scene. Based on established deep learning models, the study delves into the exploration of two distinct emotion classification metrics: The Six Emotion Prediction and the Valence/Arousal/Dominance Prediction. To facilitate a comparative analysis, a preliminary study an a following survey is deployed. The preliminary study confirms a significant difference in the generated MIDI data. For this reason, a survey is essential to detect the better fitting algorithm. Participants are tasked to rate the affective suitability of eight generated interval sequences to the corresponding film scenes. The Suitability is verified quantitatively using a bidirectional rating system. Both model assessments are conducted within a uniform sound design, thus ensuring unbiased conditions for evaluation. Upon a thorough examination of our extensive analysis, a preference for method A becomes increasingly evident.
DOI:10.34646/thn/ohmdok-1201