Dataset for emotion recognition research
Top 40.1% on sourcepulse
MELD is a comprehensive multimodal dataset for emotion recognition in multi-party conversations, derived from the Friends TV series. It provides text, audio, and visual modalities for over 13,000 utterances across more than 1400 dialogues, with each utterance labeled for seven emotions and three sentiment categories. This dataset is valuable for researchers and developers building advanced conversational AI systems, affective computing models, and dialogue generation tools that require nuanced understanding of emotional dynamics in group interactions.
How It Works
MELD extends the EmotionLines dataset by incorporating synchronized audio and visual data alongside text. Utterances are extracted with precise timestamps from TV show subtitles, ensuring alignment across modalities. The dataset's structure facilitates context-aware emotion recognition by capturing conversational flow and speaker interactions, aiming to improve performance over unimodal approaches.
Quick Start & Requirements
wget http://web.eecs.umich.edu/~mihalcea/downloads/MELD.Raw.tar.gz
pickle
). Specific feature extraction or model training may require deep learning frameworks (TensorFlow, PyTorch) and potentially GPU acceleration.Highlighted Details
Maintenance & Community
The project is associated with the declare-lab at the University of Michigan. Recent updates include new papers and SOTA baselines (COSMIC).
Licensing & Compatibility
The dataset itself is intended for research purposes. Specific licensing details for commercial use are not explicitly stated in the README, but the data is derived from a TV series, implying potential copyright considerations.
Limitations & Caveats
The dataset is based on a specific TV show, which may limit generalizability to other conversational contexts. Some utterances might have missing start/end times due to subtitle inconsistencies, requiring manual correction or omission.
1 year ago
1 week