File size: 377 Bytes
0b974f4
09c2bc6
48f30ae
09c2bc6
907a254
 
0b974f4
28486ce
0b974f4
48f30ae
907a254
0b974f4
907a254
 
09c2bc6
 
48f30ae
907a254
0b974f4
0e1834e
 
93439ee
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.3/flash_attn-2.7.3+cu12torch2.6cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
transformers-stream-generator
huggingface_hub
qwen-vl-utils
sentencepiece
opencv-python
torch==2.6.0
docling-core
transformers
torchvision
matplotlib
accelerate
requests
hf_xet
spaces
pillow
gradio
einops
peft
fpdf
timm
av