Datasets:
Update README.md
Browse files
README.md
CHANGED
|
@@ -165,6 +165,20 @@ configs:
|
|
| 165 |
|
| 166 |
**StreamingBench** evaluates **Multimodal Large Language Models (MLLMs)** in real-time, streaming video understanding tasks. ๐
|
| 167 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 168 |
## ๐๏ธ Overview
|
| 169 |
|
| 170 |
As MLLMs continue to advance, they remain largely focused on offline video comprehension, where all frames are pre-loaded before making queries. However, this is far from the human ability to process and respond to video streams in real-time, capturing the dynamic nature of multimedia content. To bridge this gap, **StreamingBench** introduces the first comprehensive benchmark for streaming video understanding in MLLMs.
|
|
|
|
| 165 |
|
| 166 |
**StreamingBench** evaluates **Multimodal Large Language Models (MLLMs)** in real-time, streaming video understanding tasks. ๐
|
| 167 |
|
| 168 |
+
------
|
| 169 |
+
|
| 170 |
+
[**NEW!** 2025.05.15] ๐ฅ: [Seed1.5-VL](https://github.com/ByteDance-Seed/Seed1.5-VL) achieved ALL model SOTA with a score of 82.80 on the Proactive Output.
|
| 171 |
+
|
| 172 |
+
[**NEW!** 2025.03.17] โญ: [ViSpeeker](https://arxiv.org/abs/2503.12769) achieved Open-Source SOTA with a score of 61.60 on the Omni-Source Understanding.
|
| 173 |
+
|
| 174 |
+
[**NEW!** 2025.01.14] ๐: [MiniCPM-o 2.6](https://github.com/OpenBMB/MiniCPM-o) achieved Streaming SOTA with a score of 66.01 on the Overall benchmark.
|
| 175 |
+
|
| 176 |
+
[**NEW!** 2025.01.06] ๐: [Dispider](https://github.com/Mark12Ding/Dispider) achieved Streaming SOTA with a score of 53.12 on the Overall benchmark.
|
| 177 |
+
|
| 178 |
+
[**NEW!** 2024.12.09] ๐: [InternLM-XComposer2.5-OmniLive](https://github.com/InternLM/InternLM-XComposer) achieved 73.79 on Real-Time Visual Understanding.
|
| 179 |
+
|
| 180 |
+
------
|
| 181 |
+
|
| 182 |
## ๐๏ธ Overview
|
| 183 |
|
| 184 |
As MLLMs continue to advance, they remain largely focused on offline video comprehension, where all frames are pre-loaded before making queries. However, this is far from the human ability to process and respond to video streams in real-time, capturing the dynamic nature of multimedia content. To bridge this gap, **StreamingBench** introduces the first comprehensive benchmark for streaming video understanding in MLLMs.
|