Add pipeline tag and library name to metadata (#1)
Browse files- Add pipeline tag and library name to metadata (0f661f42dae95a4b7c4f0597dc194f419346d92c)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
---
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen2.5-Omni-7B
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
# TimeChat-Captioner: Scripting Multi-Scene Videos with Time-Aware and Structural Audio-Visual Captions
|
|
@@ -22,6 +24,7 @@ base_model:
|
|
| 22 |
|
| 23 |
- **π Project Page:** [timechat-captioner.github.io](https://timechat-captioner.github.io/)
|
| 24 |
- **π Model:** [TimeChat-Captioner (7B)](https://huggingface.co/yaolily/TimeChat-Captioner-GRPO-7B)
|
|
|
|
| 25 |
- **π Train Dataset:** [TimeChatCap-42K](https://huggingface.co/datasets/yaolily/Timechat-OmniCaptioner-42K)
|
| 26 |
- **π Benchmark:** [OmniDCBench](https://huggingface.co/datasets/yaolily/OmniDenseCap-Benchmark)
|
| 27 |
|
|
@@ -129,7 +132,8 @@ with torch.inference_mode():
|
|
| 129 |
|
| 130 |
response = processor.decode(text_ids[0][inputs.input_ids[0].size(0):], skip_special_tokens=True)
|
| 131 |
|
| 132 |
-
print("
|
|
|
|
| 133 |
print("π¬ VIDEO DESCRIPTION:")
|
| 134 |
print("="*50)
|
| 135 |
print(response)
|
|
@@ -151,4 +155,4 @@ print("="*50)
|
|
| 151 |
primaryClass={cs.CV},
|
| 152 |
url={https://arxiv.org/abs/2602.08711}
|
| 153 |
}
|
| 154 |
-
```
|
|
|
|
| 1 |
---
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen2.5-Omni-7B
|
| 4 |
+
library_name: transformers
|
| 5 |
+
pipeline_tag: video-text-to-text
|
| 6 |
---
|
| 7 |
|
| 8 |
# TimeChat-Captioner: Scripting Multi-Scene Videos with Time-Aware and Structural Audio-Visual Captions
|
|
|
|
| 24 |
|
| 25 |
- **π Project Page:** [timechat-captioner.github.io](https://timechat-captioner.github.io/)
|
| 26 |
- **π Model:** [TimeChat-Captioner (7B)](https://huggingface.co/yaolily/TimeChat-Captioner-GRPO-7B)
|
| 27 |
+
- **π» Code:** [GitHub Repository](https://github.com/yaolinli/TimeChat-Captioner)
|
| 28 |
- **π Train Dataset:** [TimeChatCap-42K](https://huggingface.co/datasets/yaolily/Timechat-OmniCaptioner-42K)
|
| 29 |
- **π Benchmark:** [OmniDCBench](https://huggingface.co/datasets/yaolily/OmniDenseCap-Benchmark)
|
| 30 |
|
|
|
|
| 132 |
|
| 133 |
response = processor.decode(text_ids[0][inputs.input_ids[0].size(0):], skip_special_tokens=True)
|
| 134 |
|
| 135 |
+
print("
|
| 136 |
+
" + "="*50)
|
| 137 |
print("π¬ VIDEO DESCRIPTION:")
|
| 138 |
print("="*50)
|
| 139 |
print(response)
|
|
|
|
| 155 |
primaryClass={cs.CV},
|
| 156 |
url={https://arxiv.org/abs/2602.08711}
|
| 157 |
}
|
| 158 |
+
```
|