tttoaster commited on
Commit
c158010
·
verified ·
1 Parent(s): e359cdf

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +119 -0
README.md ADDED
@@ -0,0 +1,119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # ARC-Hunyuan-Video-7B
2
+
3
+ <!-- [![arXiv](https://img.shields.io/badge/arXiv-2404.14396-b31b1b.svg)](https://arxiv.org/abs/2404.14396)-->
4
+
5
+ [![Demo](https://img.shields.io/badge/ARC-Demo-blue)](https://arc.tencent.com/en/ai-demos/multimodal)
6
+ [![Static Badge](https://img.shields.io/badge/Model-Huggingface-yellow)](https://huggingface.co/TencentARC/ARC-Hunyuan-Video-7B)
7
+ [![Blog](https://img.shields.io/badge/ARC-Blog-green)](https://arc.tencent.com/en/ai-demos/multimodal)
8
+
9
+ <span style="font-size:smaller;">
10
+ Please note that in our Demo, ARC-Hunyuan-Video-7B is the model consistent with the model checkpoint and the one described in the paper, while ARC-Hunyuan-Video-7B-V0 only supports video description and summarization in Chinese.
11
+ </span>
12
+
13
+ ## Introduction
14
+
15
+ We introduce **ARC-Hunyuan-Video-7B**, a powerful multimodal model designed for _understanding real-world short videos_.
16
+ Understanding user-generated videos is actually challenging due to their complex visual elements, high
17
+ information density in both visuals and audio, and fast pacing that focuses on emotional expression and viewpoint delivery.
18
+ To address this challenge, ARC-Hunyuan-Video-7B
19
+ processes visual, audio, and textual signals end-to-end for a deep, structured understanding of video through integrating and reasoning over multimodal cues.
20
+
21
+ Compared to prior arts, we introduces a new paradigm of **Structured Video Comprehension**, with capabilities including:
22
+
23
+ - **Deep Understanding of Real-World Short Videos:** ARC-Hunyuan-Video-7B excels at analyzing user-generated content from platforms like WeChat Channels and TikTok. It goes beyond surface-level descriptions to grasp the creator's intent, emotional expression, and core message by processing complex visual elements, dense audio cues, and rapid pacing.
24
+ - **Synchronized Audio-Visual Reasoning:** The synchronization of raw visual and audio signals allows our model to answer complex questions that are impossible to solve with only one modality, such as understanding humor in a skit or details in a product review.
25
+ - **Precise Temporal Awareness:** ARC-Hunyuan-Video-7B knows not just _what_ happens, but _when_ it happens. It supports multi-granularity timestamped captioning, temporal video grounding, and detailed event summarization, making it perfect for applications like video search, highlight generation, and content analysis.
26
+ - **Advanced Reasoning and Application Versatility:** Leveraging a comprehensive multi-stage training regimen including Reinforcement Learning (RL), ARC-Hunyuan-Video-7B demonstrates strong reasoning capabilities. It supports zero-shot or few-shot fine-tuning for diverse downstream applications like video tagging, recommendation, and retrieval.
27
+
28
+ The model is capable of multi-granularity timestamped video captioning and summarization, open-ended video question answering, temporal video grounding, and
29
+ video reasoning as below,
30
+
31
+ <p align="center">
32
+ <img src="https://github.com/TencentARC/ARC-Hunyuan-Video-7B/blob/master/figures/teaser.jpg?raw=true" width="90%"/>
33
+ <p>
34
+
35
+ Specifically, ARC-Hunyuan-Video-7B is built on top of the Hunyuan-7B vision-language model with the following key designs to meet the requirements of effective structured video comprehension:
36
+
37
+ - An extra audio encoder with fine-grained visual-audio synchronization for temporally aligned visual-audio inputs
38
+ - A timestamp overlay mechanism on visual frames that explicitly provides the model with temporal awareness
39
+ - Millions of real-world videos with a totally automated bootstrapped annotation pipeline
40
+ - A comprehensive training regimen based on the finding that grounding the model in objective
41
+ tasks with RL is key to unlocking high-quality, subjective understanding
42
+
43
+ <p align="center">
44
+ <img src="https://github.com/TencentARC/ARC-Hunyuan-Video-7B/blob/master/figures/method.jpg?raw=true" width="95%"/>
45
+ <p>
46
+
47
+ ## News
48
+
49
+ - 2025.07.25: We release the [model checkpoint](https://huggingface.co/TencentARC/ARC-Hunyuan-Video-7B) and inference code of ARC-Hunyuan-Video-7B including [vLLM](https://github.com/vllm-project/vllm) version.
50
+ - 2025.07.25: We release the [API service](https://arc.tencent.com/en/document/ARC-Video-7B) of ARC-Hunyuan-Video-7B, which is supported by [vLLM](https://github.com/vllm-project/vllm). We release two versions: one is V0, which only supports video description and summarization in Chinese; the other is the version consistent with the model checkpoint and the one described in the paper.
51
+
52
+ ## Usage
53
+
54
+ ### Installation
55
+
56
+ Clone the repo and install dependent packages
57
+
58
+ ```bash
59
+ git clone https://github.com/TencentARC/ARC-Hunyuan-Video-7B.git
60
+ cd ARC-Hunyuan-Video-7B
61
+ pip install -r requirements.txt
62
+ pip install git+https://github.com/liyz15/transformers.git@arc_hunyuan_video
63
+
64
+ # For vllm, please follow the instructions below,
65
+ git submodule update --init --recursive
66
+ cd model_vllm/vllm/
67
+ export SETUPTOOLS_SCM_PRETEND_VERSION="0.8.5"
68
+ wget https://wheels.vllm.ai/ed2462030f2ccc84be13d8bb2c7476c84930fb71/vllm-1.0.0.dev-cp38-abi3-manylinux1_x86_64.whl
69
+ export VLLM_PRECOMPILED_WHEEL_LOCATION=path_of_whl
70
+ pip install --editable .
71
+ # Please install corresponding package based on your python version
72
+ pip install https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.6cxx11abiFALSE-cp311-cp311-linux_x86_64.whl
73
+ ```
74
+
75
+ ### Model Weights
76
+
77
+ - Download [ARC-Hunyuan-Video-7B](https://huggingface.co/TencentARC/ARC-Hunyuan-Video-7B) including ViT and LLM and the original [whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) .
78
+
79
+ ### Inference
80
+
81
+ #### Inference without vllm
82
+
83
+ ```bash
84
+ cd ARC-Hunyuan-Video-7B
85
+ python3 video_inference.py
86
+ ```
87
+
88
+ #### Inference with vllm
89
+
90
+ ```bash
91
+ cd ARC-Hunyuan-Video-7B
92
+ python3 video_inference_vllm.py
93
+ ```
94
+
95
+ ## API service
96
+
97
+ We also provide access to the model via API, which is supported by [vLLM](https://github.com/vllm-project/vllm). For details, please refer to the [documentation](https://arc.tencent.com/en/document/ARC-Video-7B).
98
+
99
+ We release two versions: one is V0, which only supports video description and summarization in Chinese; the other is the version consistent with the model checkpoint and the one described in the paper, which is capable of multi-granularity timestamped video captioning and summarization, open-ended video question answering, temporal video grounding, and video reasoning ( It supports Chinese and English videos and particularly excels at Chinese).
100
+
101
+ If you only need to understand and summarize short Chinese videos, we recommend using the V0 version
102
+
103
+ ## Future Work
104
+
105
+ We observe that incorporating generic video datasets during training may inadvertently compromise the model's capacity for real-world video understanding, potentially due to domain shift or noise introduced by non-real-world samples. To address this limitation, we plan to develop a dedicated model trained exclusively on rigorously curated real-world video data.
106
+
107
+ <!-- ## Citation
108
+
109
+ If you find the work helpful, please consider citing:
110
+
111
+ ```bash
112
+ @article{ge2024seed,
113
+ title={SEED-X: Multimodal Models with Unified Multi-granularity Comprehension and Generation},
114
+ author={Ge, Yuying and Zhao, Sijie and Zhu, Jinguo and Ge, Yixiao and Yi, Kun and Song, Lin and Li, Chen and Ding, Xiaohan and Shan, Ying},
115
+ journal={arXiv preprint arXiv:2404.14396},
116
+ year={2024}
117
+ }
118
+ ```
119
+ -->