dotvignesh commited on
Commit
2abb939
·
verified ·
1 Parent(s): f35ae37

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Perry-7B
2
+
3
+ A generalist reasoning LLM trained on synthetic chain-of-thought traces over STEM data. Led as a research project at Anna University (Sep 2023) — before reasoning-focused models became mainstream.
4
+
5
+ ## Overview
6
+
7
+ Perry is a fine-tuned LLaMA model designed to improve reasoning capabilities through synthetic CoT supervision. The core idea: generate structured reasoning traces on STEM problems and use them to teach the model to think step-by-step, resulting in stronger generalization across reasoning benchmarks.
8
+
9
+ Models were trained at 7B and 13B scales using compute-efficient methods.
10
+
11
+ ## Results
12
+
13
+ Improvements over baselines (as of Sep 2023):
14
+
15
+ | Benchmark | Improvement |
16
+ |-----------|-------------|
17
+ | Winogrande | +4% |
18
+ | ARC-Challenge | +6% |
19
+
20
+ ## Usage
21
+
22
+ ```python
23
+ from transformers import AutoModelForCausalLM, AutoTokenizer
24
+
25
+ model = AutoModelForCausalLM.from_pretrained("dotvignesh/perry-7b")
26
+ tokenizer = AutoTokenizer.from_pretrained("dotvignesh/perry-7b")
27
+ ```
28
+
29
+ ## Model Details
30
+
31
+ - **Base model:** LLaMA
32
+ - **Training data:** Synthetic CoT traces on STEM datasets
33
+ - **Framework:** PyTorch / Transformers