| | --- |
| | tags: |
| | - multimodal |
| | - emotion-recognition |
| | - emotional-intelligence |
| | - video |
| | - conversation |
| | language: |
| | - en |
| | license: apache-2.0 |
| | pretty_name: EmoBench-M |
| | task_categories: |
| | - video-classification |
| | - video-text-to-text |
| | dataset_info: |
| | - config: |
| | - config_name: default |
| | --- |
| | |
| | # EmoBench-M: Benchmarking Emotional Intelligence for Multimodal Large Language Models |
| |
|
| | <p align="center"> |
| | <img src="https://raw.githubusercontent.com/Emo-gml/Emo-gml.github.io/master/emo.jpg" |
| | alt="EmoBench-M logo" |
| | width="220px" /> |
| | </p> |
| |
|
| | <p align="center"> |
| | <a href="https://arxiv.org/abs/2502.04424"> |
| | <img src="https://img.shields.io/badge/arXiv-2502.04424-b31b1b.svg" alt="arXiv"> |
| | </a> |
| | <a href="https://github.com/Emo-gml/EmoBench-M"> |
| | <img src="https://img.shields.io/badge/GitHub-Emo--gml%2FEmoBench--M-blue?logo=github" alt="GitHub"> |
| | </a> |
| | </p> |
| | |
| |
|
| | ## Dataset Description |
| |
|
| | **EmoBench-M** is a comprehensive benchmark designed to evaluate the Emotional Intelligence (EI) of Multimodal Large Language Models (MLLMs). It provides a challenging testbed for assessing a model's ability to understand and interpret human emotions from video, a critical step towards developing more empathetic and human-like AI systems. |
| |
|
| | The dataset consists of video clips featuring individuals expressing various emotions. Each video is paired with a conversational prompt that asks the model to determine the emotion conveyed. This structure pushes models to integrate visual, auditory, and textual information to make a correct assessment, moving beyond simple text-based sentiment analysis. |
| | If you find this Dataset helpful, feel free to ⭐ it! [EmoBench-M](https://github.com/Emo-gml/EmoBench-M). |
| |
|
| | ## 📦 Dataset |
| |
|
| | Each JSON file contains conversation-style prompts and labels aligned with the corresponding video clips. The structure looks like: |
| |
|
| | ```json |
| | [ |
| | { |
| | "id": "0", |
| | "video": "videos/ch-simsv2s/aqgy4_0004/00023.mp4", |
| | "conversations": [ |
| | { |
| | "from": "human", |
| | "value": "<video>\nThe person in the video says: ... Determine the emotion conveyed..." |
| | }, |
| | { |
| | "from": "gpt", |
| | "value": "negative" |
| | } |
| | ] |
| | } |
| | ] |
| | ``` |
| | ### 📁 Dataset Structure |
| | ```bash |
| | EmoBench-M/ |
| | ├── benchmark_json/ # JSON files containing metadata and annotations for each dataset |
| | │ ├── FGMSA.json # Test instructions for the FGMSA dataset |
| | │ ├── MC-EIU.json # 500-sample test set for the MC-EIU dataset |
| | │ ├── MELD.json # Test instructions for the MELD dataset |
| | │ ├── MOSEI.json # 500-sample test set for the MOSEI dataset |
| | │ ├── MOSI.json # 500-sample test set for the MOSI dataset |
| | │ ├── MUSTARD.json # 500-sample test set for the MUSTARD dataset |
| | │ ├── RAVDSS_song.json # 500-sample test set for the RAVDSS song subset |
| | │ ├── RAVDSS_speech.json # 500-sample test set for the RAVDSS speech subset |
| | │ ├── SIMS.json # 500-sample test set for the SIMS dataset |
| | │ ├── ch-simsv2s.json # 500-sample test set for the Chinese SIMS v2s dataset |
| | │ ├── funny.json # Test instructions for the UR-FUNNY dataset |
| | │ ├── mer2023.json # Test instructions for the MER2023 dataset |
| | │ └── smile.json # Test data for the SMILE dataset |
| | └── dataset/ # Corresponding video files for each dataset |
| | ├── FGMSA/ |
| | │ └── videos/ |
| | │ └── FGMSA/ # Video files for the FGMSA dataset |
| | ├── MC-EIU/ |
| | │ └── videos/ |
| | │ └── MC-EIU/ # Video files for the MC-EIU dataset |
| | ├── MELD/ |
| | │ └── videos/ |
| | │ └── MELD/ # Video files for the MELD dataset |
| | ├── MOSEI/ |
| | │ └── videos/ |
| | │ └── MOSEI/ # Video files for the MOSEI dataset |
| | ├── MOSI/ |
| | │ └── videos/ |
| | │ └── MOSI/ # Video files for the MOSI dataset |
| | ├── MUSTARD/ |
| | │ └── videos/ |
| | │ └── MUSTARD/ # Video files for the MUSTARD dataset |
| | ├── RAVDSS_song/ |
| | │ └── videos/ |
| | │ └── RAVDSS/ # Video files for the RAVDSS song subset |
| | ├── RAVDSS_speech/ |
| | │ └── videos/ |
| | │ └── RAVDSS/ # Video files for the RAVDSS speech subset |
| | ├── SIMS_test/ |
| | │ └── videos/ |
| | │ └── SIMS/ # Video files for the SIMS dataset |
| | ├── ch-simsv2s/ |
| | │ └── videos/ |
| | │ └── ch-simsv2s/ # Video files for the Chinese SIMS v2s dataset |
| | ├── funny/ |
| | │ └── videos/ |
| | │ └── UR-FUNNY/ # Video files for the UR-FUNNY dataset |
| | ├── mer2023/ |
| | │ └── videos/ |
| | │ └── MER2023/ # Video files for the MER2023 dataset |
| | └── smile/ |
| | └── videos/ |
| | └── SMILE/ # Video files for the SMILE dataset |
| | ``` |
| | 📂 Dtat Structure Overview |
| | - benchmark_json/: Contains JSON files with metadata and annotations for each dataset, including test instructions and sample information. |
| | - dataset/: Corresponding video files for each dataset, organized into subdirectories named after each dataset. |
| | |