sedrickkeh commited on
Commit
03d615c
·
verified ·
1 Parent(s): 274d459

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +95 -0
README.md ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen2.5-32B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - full
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: OpenThinker2-32B
11
+ results: []
12
+ datasets:
13
+ - open-thoughts/OpenThoughts2-1M
14
+ ---
15
+
16
+ <p align="center">
17
+ <img src="https://huggingface.co/datasets/open-thoughts/open-thoughts-114k/resolve/main/open_thoughts.png" width="50%">
18
+ </p>
19
+
20
+ # OpenThinker2-32B
21
+
22
+ This model is a fine-tuned version of [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) on the
23
+ [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
24
+
25
+ The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
26
+ This model improves upon our previous [OpenThinker-7B](https://huggingface.co/open-thoughts/OpenThinker-7B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
27
+ The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
28
+
29
+ | Model | Open Data? | Avg | AIME24 | AIME25 | AMC23 | MATH500 | GPQA-D | LCBv2 |
30
+ | ---------------- | ---------- | ---- | ------ | ------ | ----- | ------- | ------ | ----- |
31
+ | OpenThinker-32B | ✅ | 72.6 | 68.0 | 49.3 | 95.5 | 90.6 | 63.5 | 68.6 |
32
+ | OpenThinker2-32B | ✅ | 76.1 | 76.7 | 58.7 | 94.0 | 90.8 | 64.1 | 72.5 |
33
+ | R1-Distill-32B | ❌ | 74.9 | 74.7 | 50.0 | 96.5 | 90.0 | 65.8 | 72.3 |
34
+ | Light-R1-32B | ✅ | 72.9 | 74.7 | 58.0 | 96.0 | 90.4 | 62.0 | 56.0 |
35
+ | QwQ-32B | ❌ | 80.9 | 78.0 | 62.0 | 98.0 | 91.6 | 66.3 | 89.2 |
36
+
37
+
38
+ # Data
39
+
40
+ This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
41
+
42
+ That dataset was constructed by augmenting [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k) with existing datasets like [OpenR1](https://huggingface.co/open-r1), as well as additional math and code reasoning data.
43
+ We generate the additional math and code data by ablating on various question generation methodologies and sampling from the highest performing ones.
44
+ See the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) model page or our [blog post]() for additional information.
45
+
46
+
47
+ ## Intended uses & limitations
48
+
49
+ Apache 2.0 License
50
+
51
+
52
+ ## Training procedure
53
+
54
+ We used 128 4xA100 nodes to train the model for 50 hours.
55
+
56
+ ### Training hyperparameters
57
+
58
+ The following hyperparameters were used during training:
59
+ - learning_rate: 8e-05
60
+ - seed: 42
61
+ - distributed_type: multi-GPU
62
+ - num_devices: 512
63
+ - gradient_accumulation_steps: 1
64
+ - total_train_batch_size: 512
65
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
66
+ - lr_scheduler_type: cosine
67
+ - lr_scheduler_warmup_ratio: 0.1
68
+ - num_epochs: 5.0
69
+
70
+ ### Framework versions
71
+
72
+ - Transformers 4.46.1
73
+ - Pytorch 2.3.0
74
+ - Datasets 3.1.0
75
+ - Tokenizers 0.20.3
76
+
77
+ More info can be found in our repository: [https://github.com/open-thoughts/open-thoughts](https://github.com/open-thoughts/open-thoughts).
78
+
79
+ # Citation
80
+ ```
81
+ @misc{openthoughts,
82
+ author = {Team, OpenThoughts},
83
+ month = apr,
84
+ title = {{Open Thoughts}},
85
+ howpublished = {https://open-thoughts.ai},
86
+ year = {2025}
87
+ }
88
+ ```
89
+
90
+ # Links
91
+ - 📊 [OpenThought2 and OpenThinker2 Blog Post]()
92
+ - 💻 [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
93
+ - 🧠 [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
94
+ - 🤖 [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
95
+ - 🤖 [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.