Korean Drama Scene Transcript Dataset for Emotion Recognition in Conversations
Understanding emotions in conversation is a challenging task as the sentences often have an implied meaning which is not generally understood in isolation. Efficient use of contextual information is important for emotion recognition in conversations. Many of the published datasets provide contextual...
Gespeichert in:
Veröffentlicht in: | IEEE access 2022, Vol.10, p.1-1 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Understanding emotions in conversation is a challenging task as the sentences often have an implied meaning which is not generally understood in isolation. Efficient use of contextual information is important for emotion recognition in conversations. Many of the published datasets provide contextual information for situations such as text-based online messaging, chatbots, and movie dialogues. However, such dialogue-based datasets are collected by selecting the ideal conversational situations and thus do not include many variations in dialogue length and number of participants. Therefore, such datasets may not be applicable for emotion recognition in text-based movie transcripts, where scenes contain variations in the number of speakers and length of the spoken sentences. We present a conversation dataset based on the Korean TV show transcripts for analysis of the emotions in presence of scene context. Korean Drama Scene Transcript dataset for Emotion Recognition (KD-EmoR) is a text-based conversation dataset. We analyze three classes of complex emotions namely euphoria, dysphoria, and neutral in the scenes of TV Drama to build a publicly available dataset for further research. We developed a context-aware deep learning model to classify the emotions utilizing speaker-level context and scene context and achieved an F1 score of 0.63 on the proposed dataset. |
---|---|
ISSN: | 2169-3536 2169-3536 |
DOI: | 10.1109/ACCESS.2022.3221408 |