yeliudev commited on
Commit
cb43899
·
verified ·
1 Parent(s): 4d9fed3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -3
README.md CHANGED
@@ -1,3 +1,41 @@
1
- ---
2
- license: bsd-3-clause
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: bsd-3-clause
3
+ pipeline_tag: video-text-to-text
4
+ ---
5
+
6
+ # VideoMind-7B
7
+
8
+ <div style="display: flex; gap: 5px;">
9
+ <a href="https://arxiv.org/abs/2503.13444" target="_blank"><img src="https://img.shields.io/badge/arXiv-2503.13444-red"></a>
10
+ <a href="https://videomind.github.io/" target="_blank"><img src="https://img.shields.io/badge/Project-Page-brightgreen"></a>
11
+ <a href="https://github.com/yeliudev/VideoMind/blob/main/README.md" target="_blank"><img src="https://img.shields.io/badge/License-BSD--3--Clause-purple"></a>
12
+ <a href="https://huggingface.co/spaces/yeliudev/VideoMind-2B" target="_blank" style="margin: 0;"><img src="https://huggingface.co/datasets/huggingface/badges/resolve/main/open-in-hf-spaces-sm-dark.svg"></a>
13
+ <a href="https://github.com/yeliudev/VideoMind" target="_blank"><img src="https://img.shields.io/github/stars/yeliudev/VideoMind"></a>
14
+ </div>
15
+
16
+ VideoMind is a multi-modal agent framework that enhances video reasoning by emulating *human-like* processes, such as *breaking down tasks*, *localizing and verifying moments*, and *synthesizing answers*.
17
+
18
+ ## 🔖 Model Details
19
+
20
+ ### Model Description
21
+
22
+ - **Model type:** Multi-modal Large Language Model
23
+ - **Language(s):** English
24
+ - **License:** BSD-3-Clause
25
+
26
+ ### More Details
27
+
28
+ Please refer to our [GitHub Repository](https://github.com/yeliudev/VideoMind) for more details about this model.
29
+
30
+ ## 📖 Citation
31
+
32
+ Please kindly cite our paper if you find this project helpful.
33
+
34
+ ```
35
+ @article{liu2025videomind,
36
+ title={VideoMind: A Chain-of-LoRA Agent for Long Video Reasoning},
37
+ author={Liu, Ye and Lin, Kevin Qinghong and Chen, Chang Wen and Shou, Mike Zheng},
38
+ journal={arXiv preprint arXiv:2503.13444},
39
+ year={2025}
40
+ }
41
+ ```