aeb56 commited on
Commit
9bb160e
·
1 Parent(s): a951334

Update requirements.txt with comprehensive dependencies for Kimi model

Browse files
Files changed (1) hide show
  1. requirements.txt +59 -11
requirements.txt CHANGED
@@ -1,12 +1,60 @@
1
- torch>=2.1.0
2
- transformers>=4.40.0
3
- peft>=0.10.0
4
- accelerate>=0.27.0
5
- bitsandbytes>=0.42.0
6
- gradio>=4.19.0
7
- huggingface-hub>=0.20.0
8
- sentencepiece>=0.1.99
9
- protobuf>=3.20.0
10
- safetensors>=0.4.0
11
- scipy>=1.10.0
12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Core ML Libraries
2
+ transformers>=4.56.0 # Required by Kimi model (has assertion check)
3
+ accelerate>=0.34.2 # Compatible with latest transformers
4
+ peft>=0.13.2 # Latest stable
5
+ bitsandbytes>=0.45.1 # Compatible with triton 3.x (0.45.0 has triton.ops issue)
6
+ sentencepiece==0.2.0
7
+ protobuf==5.29.2 # Updated for compatibility
 
 
 
 
8
 
9
+ # Training & Optimization
10
+ deepspeed==0.16.3 # Compatible with torch 2.5, more stable than 0.18.x
11
+ triton>=3.2.0 # Required by fla for optimal performance
12
+ scipy==1.14.1 # Updated
13
+ scikit-learn==1.6.0 # Updated
14
+ ninja==1.11.1.1
15
+
16
+ # Data Processing
17
+ datasets>=3.2.0 # Updated for compatibility
18
+ tokenizers>=0.21.0 # Compatible with latest transformers
19
+ pandas==2.2.3
20
+ numpy==1.26.4 # Keep for stability (2.x has breaking changes)
21
+
22
+ # Monitoring & Logging
23
+ wandb==0.19.1 # Updated
24
+ tensorboard==2.18.0
25
+ tqdm==4.67.1 # Updated
26
+ psutil==6.1.1 # Updated
27
+ pynvml==11.5.3 # Updated
28
+
29
+ # Evaluation
30
+ rouge-score==0.1.2
31
+ sacrebleu==2.4.3 # Updated
32
+ bert-score==0.3.13
33
+
34
+ # Utilities
35
+ pyyaml==6.0.2
36
+ python-dotenv==1.0.1
37
+ huggingface-hub>=0.34.0 # Required by transformers >=4.56.0
38
+ safetensors==0.4.5
39
+ tiktoken==0.8.0 # Updated
40
+ hf_transfer==0.1.8 # Updated
41
+
42
+ # Kimi / Flash Linear Attention runtime (requires torch>=2.5)
43
+ # Install from git to get latest version with fla.layers module
44
+ git+https://github.com/sustcsonglin/flash-linear-attention.git@main
45
+
46
+ # Required by Kimi tokenizer (tiktoken BPE loader)
47
+ blobfile==3.0.0 # Updated
48
+
49
+ # Web UI for HF Space
50
+ gradio==4.44.1 # Web interface to keep Space alive
51
+
52
+ # API (optional - not used with Gradio)
53
+ # fastapi==0.115.6
54
+ # uvicorn[standard]==0.34.0
55
+ # python-multipart==0.0.20
56
+
57
+ # Development
58
+ pytest==8.3.4 # Updated
59
+ black==24.10.0 # Updated
60
+ flake8==7.1.1