File size: 14,475 Bytes
8d08eab
 
 
 
c1d1b43
 
8d08eab
 
c1d1b43
8d08eab
 
c1d1b43
8d08eab
 
 
a9ee241
274ff39
8d08eab
 
 
 
 
 
 
 
274ff39
 
a9ee241
274ff39
240c6af
274ff39
be76504
 
274ff39
 
 
 
 
 
 
 
 
 
 
 
c2c113c
274ff39
7c9fe0a
 
 
 
 
 
 
 
 
 
 
 
 
 
274ff39
 
1055d77
 
a9ee241
1055d77
722e0c3
1055d77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
240c6af
1055d77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7c9fe0a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1055d77
 
 
 
 
 
 
 
 
 
c1d1b43
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
---
license: cc-by-4.0
task_categories:
- zero-shot-classification
- text-classification
- text-generation
language:
- en
- zh
size_categories:
- 10K<n<100K
pretty_name: MMLA
---
# Can Large Language Models Help Multimodal Language Analysis? MMLA: A Comprehensive Benchmark

## 1. Introduction

MMLA is the first comprehensive multimodal language analysis benchmark for evaluating foundation models. It has the following features:

- **Large** Scale: 61K+ multimodal samples.
- **Various** Sources: 9 datasets.
- **Three** Modalities: text, video, and audio
- Both **Acting** and **Real-world** Scenarios: films, TV series, YouTube, Vimeo, Bilibili, TED, improvised scripts, etc.
- **Six** Core Dimensions in Multimodal Language Analysis: intent, emotion, sentiment, dialogue act, speaking style, and communication behavior.

We also build baselines with three evaluation methods (i.e., zero-shot inference, supervised fine-tuning, and instruction tuning) on 8 mainstream foundation models (i.e., 5 MLLMs (Qwen2-VL, VideoLLaMA2, LLaVA-Video, LLaVA-OV, MiniCPM-V-2.6), 3 LLMs (InternLM2.5, Qwen2, LLaMA3). More details can refer to our [paper](https://arxiv.org/abs/2504.16427).

## 2. Datasets

#### 2.1 Statistics 

Dataset statistics for each dimension in the MMLA benchmark. #C, #U, #Train, #Val, and #Test represent the number of label classes, utterances, training, validation, and testing samples, respectively. avg. and max. refer to the average and maximum lengths.  
| **Dimensions**            | **Datasets**          | **#C** | **#U**  | **#Train** | **#Val** | **#Test** | **Video Hours** | **Source**        | **#Video Length (avg. / max.)** | **#Text Length (avg. / max.)** | **Language** |
|---------------------------|-----------------------|--------|---------|------------|----------|-----------|-----------------|-------------------|---------------------------------|---------------------------------|--------------|
| **Intent**                | MIntRec               | 20     | 2,224   | 1,334      | 445      | 445       | 1.5             | TV series         | 2.4 / 9.6                     | 7.6 / 27.0                     | **English** |
|                           | MIntRec2.0            | 30     | 9,304   | 6,165      | 1,106    | 2,033     | 7.5             | TV series         | 2.9 / 19.9                    | 8.5 / 46.0                     |              |
| **Dialogue Act**          | MELD                  | 12     | 9,989   | 6,992      | 999      | 1,998     | 8.8             | TV series         | 3.2 / 41.1                    | 8.6 / 72.0                     | **English** |
|                           | IEMOCAP               | 12     | 9,416   | 6,590      | 942      | 1,884     | 11.7            | Improvised scripts | 4.5 / 34.2                    | 12.4 / 106.0                    |              |
| **Emotion**               | MELD                  | 7      | 13,708  | 9,989      | 1,109    | 2,610     | 12.2            | TV series         | 3.2 / 305.0                   | 8.7 / 72.0                     | **English** |
|                           | IEMOCAP               | 6      | 7,532   | 5,237      | 521      | 1,622     | 9.6             | Improvised scripts | 4.6 / 34.2                    | 12.8 / 106.0                    |              |
| **Sentiment**             | MOSI                  | 2      | 2,199   | 1,284      | 229      | 686       | 2.6             | Youtube           | 4.3 / 52.5                    | 12.5 / 114.0                    | **English** |
|                           | CH-SIMS v2.0          | 3      | 4,403   | 2,722      | 647      | 1,034     | 4.3             | TV series, films  | 3.6 / 42.7                    | 1.8 / 7.0                      | **Mandarin**|
| **Speaking Style**        | UR-FUNNY-v2           | 2      | 9,586   | 7,612      | 980      | 994       | 12.9            | TED               | 4.8 / 325.7                   | 16.3 / 126.0                    | **English** |
|                           | MUStARD               | 2      | 690     | 414        | 138      | 138       | 1.0             | TV series         | 5.2 / 20.0                    | 13.1 / 68.0                     |              |
| **Communication Behavior**| Anno-MI (client)      | 3      | 4,713   | 3,123      | 461      | 1,128     | 10.8            | YouTube & Vimeo   | 8.2 / 600.0                   | 16.3 / 266.0                    | **English** |
|                           | Anno-MI (therapist)   | 4      | 4,773   | 3,161      | 472      | 1,139     | 12.1            |                   | 9.1 / 1316.1                  | 17.9 / 205.0                    |              |


#### 2.2 Collection Timeline
- **MIntRec**: Released in 2022/10.
- **MIntRec2.0**: Released in 2024/01.
- **MELD**: Collected from TV series (released in 2019/05).
- **UR-FUNNY-v2**: Collected from publicly available TED talks (released in 2019/11).
- **MUStARD**: Collected from TV series (released in 2019/07).
- **MELD-DA**: Dialogue act annotations added to MELD in 2020/07.
- **IEMOCAP-DA**: Dialogue act annotations added to IEMOCAP (released in 2020/07).
- **MOSI**: Collected from YouTube opinion videos (released in 2016/06).
- **IEMOCAP**: Collected from scripted improvisational acting (released in 2008/12).
- **Anno-MI**: Collected from publicly available YouTube and Vimeo videos (released in 2023/03).
- 
#### 2.3 License

This benchmark uses nine datasets, each of which is employed strictly in accordance with its official license and exclusively for academic research purposes. We fully respect the datasets’ copyright policies, license requirements, and ethical standards. For those datasets whose licenses explicitly permit redistribution, we release the original video data (e.g., [MIntRec](https://github.com/thuiar/MIntRec), [MIntRec2.0](https://github.com/thuiar/MIntRec2.0), [MELD](https://github.com/declare-lab/MELD), [UR-FUNNY-v2](https://github.com/ROC-HCI/UR-FUNNY), [MUStARD](https://github.com/soujanyaporia/MUStARD), [MELD-DA](https://github.com/sahatulika15/EMOTyDA), [CH-SIMS v2.0](https://github.com/thuiar/ch-sims-v2), and [Anno-MI](https://github.com/uccollab/AnnoMI). For datasets that restrict video redistribution, users should obtain the videos directly from their official repositories (e.g., [MOSI](https://github.com/matsuolab/CMU-MultimodalSDK), [IEMOCAP and IEMOCAP-DA](https://sail.usc.edu/iemocap). In compliance with all relevant licenses, we also provide the original textual data unchanged, together with the specific dataset splits used in our experiments. This approach ensures reproducibility and academic transparency while strictly adhering to copyright obligations and protecting the privacy of individuals featured in the videos.


## 3. LeaderBoard

#### 3.1 Rank of Zero-shot Inference

| RANK |      Models      |  ACC  | TYPE |
| :--: | :--------------: | :---: | :--: |
|  🥇   |      GPT-4o      | 52.60 | MLLM |
|  🥈   |   Qwen2-VL-72B   | 52.55 | MLLM |
|  🥉   |   LLaVA-OV-72B   | 52.44 | MLLM |
|  4   | LLaVA-Video-72B  | 51.64 | MLLM |
|  5   |  InternLM2.5-7B  | 50.28 | LLM  |
|  6   |     Qwen2-7B     | 48.45 | LLM  |
|  7   |   Qwen2-VL-7B    | 47.12 | MLLM |
|  8   |    Llama3-8B     | 44.06 | LLM  |
|  9   |  LLaVA-Video-7B  | 43.32 | MLLM |
|  10  |  VideoLLaMA2-7B  | 42.82 | MLLM |
|  11  |   LLaVA-OV-7B    | 40.65 | MLLM |
|  12  |    Qwen2-1.5B    | 40.61 | LLM  |
|  13  | MiniCPM-V-2.6-8B | 37.03 | MLLM |
|  14  |    Qwen2-0.5B    | 22.14 | LLM  |

#### 3.2 Rank of Supervised Fine-tuning (SFT) and Instruction Tuning (IT)

| Rank |         Models         |  ACC  | Type |
| :--: | :--------------------: | :---: | :--: |
|  🥇   |   Qwen2-VL-72B (SFT)   | 69.18 | MLLM |
|  🥈   | MiniCPM-V-2.6-8B (SFT) | 68.88 | MLLM |
|  🥉   |  LLaVA-Video-72B (IT)  | 68.87 | MLLM |
|  4   |   LLaVA-ov-72B (SFT)   | 68.67 | MLLM |
|  5   |   Qwen2-VL-72B (IT)    | 68.64 | MLLM |
|  6   | LLaVA-Video-72B (SFT)  | 68.44 | MLLM |
|  7   |  VideoLLaMA2-7B (SFT)  | 68.30 | MLLM |
|  8   |   Qwen2-VL-7B (SFT)    | 67.60 | MLLM |
|  9   |   LLaVA-ov-7B (SFT)    | 67.54 | MLLM |
|  10  |  LLaVA-Video-7B (SFT)  | 67.47 | MLLM |
|  11  |    Qwen2-VL-7B (IT)    | 67.34 | MLLM |
|  12  | MiniCPM-V-2.6-8B (IT)  | 67.25 | MLLM |
|  13  |    Llama-3-8B (SFT)    | 66.18 | LLM  |
|  14  |     Qwen2-7B (SFT)     | 66.15 | LLM  |
|  15  | Internlm-2.5-7B (SFT)  | 65.72 | LLM  |
|  16  |     Qwen-2-7B (IT)     | 64.58 | LLM  |
|  17  |  Internlm-2.5-7B (IT)  | 64.41 | LLM  |
|  18  |    Llama-3-8B (IT)     | 64.16 | LLM  |
|  19  |    Qwen2-1.5B (SFT)    | 64.00 | LLM  |
|  20  |    Qwen2-0.5B (SFT)    | 62.80 | LLM  |


## 4. Data Integrity

All files included in the MMLA benchmark are verified using SHA-256 checksums. Please ensure the integrity of the files using the following checksums:

| File Path | SHA256 Hash |
|-----------|-------------|
| `/MMLA-Datasets/AnnoMi-client/test.tsv` | `d555c7131bc54cb61424d421c7a3ec117fa5587c1d4027dd8501321a5d1abc09` |
| `/MMLA-Datasets/AnnoMi-client/dev.tsv` | `4695dc4e1c360cac53ecd0386b82d10fdda9414ad1d559c0a9491a8981657acd` |
| `/MMLA-Datasets/AnnoMi-client/train.tsv` | `8e1104e7d4e42952d0e615c22ee7ea08c03d9b7d07807ba6f4fd4b41d08fed89` |
| `/MMLA-Datasets/AnnoMi-client/AnnoMI-client_video.tar.gz` | `597d9b8c1a701a89c3f6b18e4a451c21b6699a670a83225b7bce5212f5abdfe0` |
| `/MMLA-Datasets/AnnoMi-therapist/dev.tsv` | `bde3ae0e4f16e2249ac94245802b1e5053df3c9d4864f8a889347fe492364767` |
| `/MMLA-Datasets/AnnoMi-therapist/test.tsv` | `0ef6ceeba7dfff9f3201b263aecdb6636b6dd39c5eec220c91a328b5dd23e9d5` |
| `/MMLA-Datasets/AnnoMi-therapist/train.tsv` | `fd0a4741bd3fb32014318f0bd0fbc464a87a9e267163fcac9618707fedca12b2` |
| `/MMLA-Datasets/AnnoMi-therapist/AnnoMi-therapist_video.tar.gz` | `767ce57ad55078001cdd616d642f78d3b0433d9ebcbc14db1608408a54c9fa10` |
| `/MMLA-Datasets/CH-SIMSv2.0/test.tsv` | `40afae5245b1060e8bb5162e8cc4f17f294a43b51a9e01e5bbd64d1f5ebcb6d7` |
| `/MMLA-Datasets/CH-SIMSv2.0/dev.tsv` | `47dfac9ca8d77868ed644b8cd9536fa403f9d6f81e26796cd882e39d2cc14608` |
| `/MMLA-Datasets/CH-SIMSv2.0/train.tsv` | `96350a9e35d62dc63035256e09f033f84aa670f6bf1c06e38daef85d39bde7d7` |
| `/MMLA-Datasets/CH-SIMSv2.0/Ch-simsv2_video.tar.gz` | `e2817c4841a74f9e73eed6cf3196442ff0245f999bdfc5f975dcf18e66348f1e` |
| `/MMLA-Datasets/IEMOCAP-DA/dev.tsv` | `67d357fee50c9b009f9cdc81738e1f45562e0a7f193f6f100320e1881d2b2c8c` |
| `/MMLA-Datasets/IEMOCAP-DA/test.tsv` | `050d27887bec3714f8f0c323594c3c287fa9a5c006f94de0fa09565ba0251773` |
| `/MMLA-Datasets/IEMOCAP-DA/train.tsv` | `823b37fa045aa6aad694d94ad134e23b92491cd6c5d742ed6e9d9456b433608b` |
| `/MMLA-Datasets/IEMOCAP/dev.tsv` | `b6b0bbe1f49dc1f20c4121ac8f943b2d85722c95bb0988946282a496c0c1094d` |
| `/MMLA-Datasets/IEMOCAP/test.tsv` | `7ab10d9c126e037e8c6be1ddf6487d57e9132b2e238286a6a9cccce029760581` |
| `/MMLA-Datasets/IEMOCAP/train.tsv` | `a0017547086721147ed1191e8b7d5da42f795c4070687cffcff001d8827b81d8` |
| `/MMLA-Datasets/MELD-DA/test.tsv` | `b25f4396f30a8d591224ec8074cc4ebfd5727f22fa816ab46cdb455dc22ee854` |
| `/MMLA-Datasets/MELD-DA/dev.tsv` | `4fcc28d139ac933df8e8a288f2d17e010d5e013c70722485a834a7b843536351` |
| `/MMLA-Datasets/MELD-DA/train.tsv` | `045642a0abaa9d9d9ea5f7ade96a09dd856311c9a375dea1839616688240ec71` |
| `/MMLA-Datasets/MELD-DA/MELD-DA_video.tar.gz` | `92154bb5d2cf9d8dc229d5fe7ce65519ee7525487f4f42ca7acdf79e48c69707` |
| `/MMLA-Datasets/MELD/dev.tsv` | `ce677f8162ce901e0cc26f531f1786620cac40b7507fa34664537dadc407d256` |
| `/MMLA-Datasets/MELD/test.tsv` | `ee0e0a35a8ae73b522f359039cea34e92d0e13283f5f01c4f29795b439a92a69` |
| `/MMLA-Datasets/MELD/train.tsv` | `063ac8accce2e0da3b45e9cdb077c5374a4cf08f6d62db41438e6e0c52981287` |
| `/MMLA-Datasets/MELD/MELD_video.tar.gz` | `6ce66e5e0d3054aeaf2f5857106360f3b94c37e099bf2e2b17bc1304ef79361b` |
| `/MMLA-Datasets/MIntRec/dev.tsv` | `629ab568ec3e1343c83d76b43d7398f7580361370d09162065a6bb1883f2fe9a` |
| `/MMLA-Datasets/MIntRec/test.tsv` | `adffdc8f061878ad560ee0e0046ba32e6bc9e0332d9e09094cfce0b755fcc2a9` |
| `/MMLA-Datasets/MIntRec/train.tsv` | `c1bec2ff06712063c7399264d7c06f4cdc125084314e6fa8bdfd94d3f0b42332` |
| `/MMLA-Datasets/MIntRec/MIntRec_video.tar.gz` | `a756b6ad5f851773b3ae4621e3aa5c33a662bde80b239a6815a8541c30fc6411` |
| `/MMLA-Datasets/MIntRec2.0/dev.tsv` | `f2f69111d0bd8c26681db0a613a0112f466c667d56a79949ce17ccadd1e6ae37` |
| `/MMLA-Datasets/MIntRec2.0/test.tsv` | `6aa650afbaf40256afdbb546a9f7253511f3fe8d791a9acc7b6829824455a6ed` |
| `/MMLA-Datasets/MIntRec2.0/train.tsv` | `e8b8767bd9a4de5833475db2438e63390c9674041a7b8ea39183a74fa4b624ef` |
| `/MMLA-Datasets/MIntRec2.0/MIntRec2.0_video.tar.gz` | `78bd9ab4a0f9e5768ed2a094524165ecc51926e210a4701a9548d036a68d5e29` |
| `/MMLA-Datasets/MOSI/dev.tsv` | `bd8ccded8dacb9cb7d37743f54c7e4c7bef391069b67b55c7e0cf4626fadee5f` |
| `/MMLA-Datasets/MOSI/test.tsv` | `c480fc2cb444d215e5ba3433452db546fd8e638d332ee0f03278158b69375eca` |
| `/MMLA-Datasets/MOSI/train.tsv` | `f1afe6018ae0b0ab8833da6934c0847f480412ed11c9c22e204a01e8cf75971b` |
| `/MMLA-Datasets/MUStARD/MUStARD_video.tar.gz` | `8bd863c7ab4c29a710aa3edc0f560361275830a1e98ec41908d51c43e08647c1` |
| `/MMLA-Datasets/MUStARD/dev.tsv` | `45477e0bda84c3d45ff197734b3943fc30e9f89c0d0cb8c272f0c10d31ee5474` |
| `/MMLA-Datasets/MUStARD/test.tsv` | `ae248884d42d690700b6ce9930bb12827cd0fbcae200c43aace5a90003ad99e5` |
| `/MMLA-Datasets/MUStARD/train.tsv` | `4292e07a087978a08552268b6c8405d897ee855af495e7e58ee99863e705eb43` |
| `/MMLA-Datasets/UR-FUNNY-v2/dev.tsv` | `a82f758ef5d2a65bc41e09e24a616d4654c1565e851cd42c71a575b09282a2d2` |
| `/MMLA-Datasets/UR-FUNNY-v2/test.tsv` | `6cb9dee9fd55545f46cd079ecb7541981d4c19a76c0ce79d7d874fe73703b63a` |
| `/MMLA-Datasets/UR-FUNNY-v2/train.tsv` | `8eb91657faa19a2d53cc930c810d2fa3abd8e365c49d27fa6feb68cd95f40fb4` |
| `/MMLA-Datasets/UR-FUNNY-v2/UR-FUNNYv2_video.tar.gz` | `e5a3962985c8ead5f593db69ab77a9d6702895768bb5871fe8764406358f8cae` |




## 5. Acknowledgements 

For more details, please refer to our [Github repo](https://github.com/thuiar/MMLA). If our work is helpful to your research, please consider citing the following paper: 

```
@article{zhang2025mmla,
  author={Zhang, Hanlei and Li, Zhuohang and Zhu, Yeshuang and Xu, Hua and Wang, Peiwu and Zhu, Haige and Zhou, Jie and Zhang, Jinchao},
  title={Can Large Language Models Help Multimodal Language Analysis? MMLA: A Comprehensive Benchmark},
  year={2025},
  journal={arXiv preprint arXiv:2504.16427},
}
```