aphoticshaman commited on
Commit
b39d100
·
verified ·
1 Parent(s): bef9b8f

Add comprehensive model card

Browse files
Files changed (1) hide show
  1. README.md +50 -5
README.md CHANGED
@@ -1,12 +1,57 @@
1
  ---
2
- base_model: Qwen/Qwen2.5-72B-Instruct
3
- library_name: peft
4
  license: apache-2.0
 
5
  tags:
6
- - elle
7
  - math
 
 
8
  - lora
9
- - qwen
 
 
 
 
 
10
  ---
 
11
  # Elle-72B-Math-v2
12
- Fine-tuned on NuminaMath-CoT for mathematical reasoning.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
 
2
  license: apache-2.0
3
+ base_model: Qwen/Qwen2.5-72B-Instruct
4
  tags:
 
5
  - math
6
+ - reasoning
7
+ - qwen2
8
  - lora
9
+ - peft
10
+ - numina
11
+ library_name: peft
12
+ pipeline_tag: text-generation
13
+ datasets:
14
+ - AI-MO/NuminaMath-CoT
15
  ---
16
+
17
  # Elle-72B-Math-v2
18
+
19
+ ## Model Description
20
+
21
+ Elle-72B-Math-v2 is a LoRA adapter fine-tuned on [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) for mathematical reasoning using the NuminaMath-CoT dataset.
22
+
23
+ ## Model Details
24
+
25
+ - **Base Model**: Qwen/Qwen2.5-72B-Instruct
26
+ - **Adapter Type**: LoRA
27
+ - **LoRA Rank**: 64
28
+ - **LoRA Alpha**: 128
29
+ - **Target Modules**: All linear layers
30
+ - **Training Data**: NuminaMath-CoT
31
+
32
+ ## Training
33
+
34
+ Fine-tuned using:
35
+ - Chain-of-thought mathematical reasoning examples
36
+ - Step-by-step problem decomposition
37
+ - Multiple solution strategies (algebraic, numerical, symbolic)
38
+
39
+ ## Usage
40
+
41
+ ```python
42
+ from peft import PeftModel
43
+ from transformers import AutoModelForCausalLM, AutoTokenizer
44
+
45
+ base_model = AutoModelForCausalLM.from_pretrained(
46
+ "Qwen/Qwen2.5-72B-Instruct",
47
+ torch_dtype="auto",
48
+ device_map="auto",
49
+ trust_remote_code=True
50
+ )
51
+ model = PeftModel.from_pretrained(base_model, "aphoticshaman/elle-72b-math-v2")
52
+ tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-72B-Instruct")
53
+ ```
54
+
55
+ ## License
56
+
57
+ Apache 2.0