File size: 1,288 Bytes
bef9b8f
 
b39d100
bef9b8f
 
b39d100
 
bef9b8f
b39d100
 
 
 
 
 
bef9b8f
b39d100
bef9b8f
b39d100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
---
license: apache-2.0
base_model: Qwen/Qwen2.5-72B-Instruct
tags:
  - math
  - reasoning
  - qwen2
  - lora
  - peft
  - numina
library_name: peft
pipeline_tag: text-generation
datasets:
  - AI-MO/NuminaMath-CoT
---

# Elle-72B-Math-v2

## Model Description

Elle-72B-Math-v2 is a LoRA adapter fine-tuned on [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) for mathematical reasoning using the NuminaMath-CoT dataset.

## Model Details

- **Base Model**: Qwen/Qwen2.5-72B-Instruct
- **Adapter Type**: LoRA
- **LoRA Rank**: 64
- **LoRA Alpha**: 128
- **Target Modules**: All linear layers
- **Training Data**: NuminaMath-CoT

## Training

Fine-tuned using:
- Chain-of-thought mathematical reasoning examples
- Step-by-step problem decomposition
- Multiple solution strategies (algebraic, numerical, symbolic)

## Usage

```python
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer

base_model = AutoModelForCausalLM.from_pretrained(
    "Qwen/Qwen2.5-72B-Instruct",
    torch_dtype="auto",
    device_map="auto",
    trust_remote_code=True
)
model = PeftModel.from_pretrained(base_model, "aphoticshaman/elle-72b-math-v2")
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-72B-Instruct")
```

## License

Apache 2.0