Korean Drama Scene Transcript Dataset for Emotion Recognition in Conversations

Understanding emotions in conversation is a challenging task as the sentences often have an implied meaning which is not generally understood in isolation. Efficient use of contextual information is important for emotion recognition in conversations. Many of the published datasets provide contextual...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2022, Vol.10, p.1-1
Hauptverfasser: Pant, Sudarshan, Lim, Eunchae, Yang, Hyung-Jeong, Lee, Guee-Sang, Kim, Soo-Hyung, Kang, Young-Shin, Jang, Hyerim
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Understanding emotions in conversation is a challenging task as the sentences often have an implied meaning which is not generally understood in isolation. Efficient use of contextual information is important for emotion recognition in conversations. Many of the published datasets provide contextual information for situations such as text-based online messaging, chatbots, and movie dialogues. However, such dialogue-based datasets are collected by selecting the ideal conversational situations and thus do not include many variations in dialogue length and number of participants. Therefore, such datasets may not be applicable for emotion recognition in text-based movie transcripts, where scenes contain variations in the number of speakers and length of the spoken sentences. We present a conversation dataset based on the Korean TV show transcripts for analysis of the emotions in presence of scene context. Korean Drama Scene Transcript dataset for Emotion Recognition (KD-EmoR) is a text-based conversation dataset. We analyze three classes of complex emotions namely euphoria, dysphoria, and neutral in the scenes of TV Drama to build a publicly available dataset for further research. We developed a context-aware deep learning model to classify the emotions utilizing speaker-level context and scene context and achieved an F1 score of 0.63 on the proposed dataset.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2022.3221408