eltokh7 commited on
Commit
5f069f6
·
verified ·
1 Parent(s): 0dfc15a

Add model card

Browse files
Files changed (1) hide show
  1. README.md +53 -0
README.md ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: Qwen/Qwen3-14B
4
+ tags:
5
+ - code
6
+ - stata
7
+ - qwen
8
+ - 14b
9
+ - finetuned
10
+ ---
11
+
12
+ # Statacoder 14B 01
13
+
14
+ LoRA adapter for Stata code generation. Base: Qwen/Qwen3-14B
15
+
16
+ ## Model Details
17
+
18
+ - **Base Model**: Qwen/Qwen3-14B
19
+ - **Model Type**: LoRA fine-tuned for Stata code generation
20
+ - **Parameters**: 14B
21
+ - **Training**: MSML bootstrap loop
22
+
23
+ ## Usage
24
+
25
+ ### With LoRA (requires base model)
26
+ ```python
27
+ from transformers import AutoModelForCausalLM, AutoTokenizer
28
+ from peft import PeftModel
29
+
30
+ base_model = "Qwen/Qwen3-14B"
31
+ adapter_path = "eltokh7/statacoder-14b-01"
32
+
33
+ model = AutoModelForCausalLM.from_pretrained(base_model, device_map="auto")
34
+ tokenizer = AutoTokenizer.from_pretrained(base_model)
35
+ model = PeftModel.from_pretrained(model, adapter_path)
36
+
37
+ prompt = "Write a Stata program to calculate mean of variable x"
38
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
39
+ outputs = model.generate(**inputs, max_new_tokens=512)
40
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
41
+ ```
42
+
43
+ ## Training
44
+
45
+ Trained using the MSML bootstrap loop with:
46
+ - 250 Stata coding problems
47
+ - LLM-powered quality gates
48
+ - Self-repair for failed solutions
49
+ - Multi-task learning expansion
50
+
51
+ ## License
52
+
53
+ MIT