real-jiakai commited on
Commit
2958745
·
verified ·
1 Parent(s): 9b6cc61

Update README.md

Browse files

Fix training data distribution in model card

Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -21,24 +21,22 @@ A math-focused fine-tuned version of SmolLM3-3B, optimized for step-by-step math
21
 
22
  ## Highlights
23
 
24
- 🧮 **Math-First**: Trained on 15K high-quality math reasoning samples
25
  🧠 **Chain-of-Thought**: Supports `/think` mode for detailed reasoning
26
  ⚡ **Lightweight**: 3B parameters, runs on consumer GPUs
27
-
28
  ## Training Details
29
 
30
  | Parameter | Value |
31
  |-----------|-------|
32
  | Base Model | HuggingFaceTB/SmolLM3-3B |
33
  | Method | LoRA (r=16, alpha=32) |
34
- | Training Data | ~15K samples |
35
- | - OpenThoughts3 | 12,000 (math reasoning) |
36
- | - s1k | 835 (high-quality math) |
37
- | - Mixture of Thoughts | 2,000 (science reasoning) |
38
  | Epochs | 2 |
39
  | Learning Rate | 2e-4 (cosine) |
40
  | Effective Batch Size | 16 |
41
-
42
  ## Usage
43
  ```python
44
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
21
 
22
  ## Highlights
23
 
24
+ 📚 **Math-First**: Trained on ~7K high-quality math and reasoning samples
25
  🧠 **Chain-of-Thought**: Supports `/think` mode for detailed reasoning
26
  ⚡ **Lightweight**: 3B parameters, runs on consumer GPUs
 
27
  ## Training Details
28
 
29
  | Parameter | Value |
30
  |-----------|-------|
31
  | Base Model | HuggingFaceTB/SmolLM3-3B |
32
  | Method | LoRA (r=16, alpha=32) |
33
+ | Training Data | ~7K samples |
34
+ | - OpenThoughts3_1.2M_think | 5,000 (math reasoning) |
35
+ | - s1k_1.1_think | ~1,000 (high-quality math) |
36
+ | - smoltalk_everyday_convs | 1,000 (everyday reasoning) |
37
  | Epochs | 2 |
38
  | Learning Rate | 2e-4 (cosine) |
39
  | Effective Batch Size | 16 |
 
40
  ## Usage
41
  ```python
42
  from transformers import AutoModelForCausalLM, AutoTokenizer