Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Perry-7B
|
| 2 |
+
|
| 3 |
+
A generalist reasoning LLM trained on synthetic chain-of-thought traces over STEM data. Led as a research project at Anna University (Sep 2023) — before reasoning-focused models became mainstream.
|
| 4 |
+
|
| 5 |
+
## Overview
|
| 6 |
+
|
| 7 |
+
Perry is a fine-tuned LLaMA model designed to improve reasoning capabilities through synthetic CoT supervision. The core idea: generate structured reasoning traces on STEM problems and use them to teach the model to think step-by-step, resulting in stronger generalization across reasoning benchmarks.
|
| 8 |
+
|
| 9 |
+
Models were trained at 7B and 13B scales using compute-efficient methods.
|
| 10 |
+
|
| 11 |
+
## Results
|
| 12 |
+
|
| 13 |
+
Improvements over baselines (as of Sep 2023):
|
| 14 |
+
|
| 15 |
+
| Benchmark | Improvement |
|
| 16 |
+
|-----------|-------------|
|
| 17 |
+
| Winogrande | +4% |
|
| 18 |
+
| ARC-Challenge | +6% |
|
| 19 |
+
|
| 20 |
+
## Usage
|
| 21 |
+
|
| 22 |
+
```python
|
| 23 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 24 |
+
|
| 25 |
+
model = AutoModelForCausalLM.from_pretrained("dotvignesh/perry-7b")
|
| 26 |
+
tokenizer = AutoTokenizer.from_pretrained("dotvignesh/perry-7b")
|
| 27 |
+
```
|
| 28 |
+
|
| 29 |
+
## Model Details
|
| 30 |
+
|
| 31 |
+
- **Base model:** LLaMA
|
| 32 |
+
- **Training data:** Synthetic CoT traces on STEM datasets
|
| 33 |
+
- **Framework:** PyTorch / Transformers
|