aman-jaglan commited on
Commit
0dddd69
·
verified ·
1 Parent(s): bd3397c

update README.md

Browse files
Files changed (1) hide show
  1. README.md +75 -3
README.md CHANGED
@@ -1,3 +1,75 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ ---
6
+ base_model: Qwen/Qwen3-8B
7
+ tags:
8
+ - adaptive-teaching
9
+ - reinforcement-learning
10
+ - educational
11
+ datasets:
12
+ - Arc-Intelligence/Arc-ATLAS-Teach-v0
13
+ language:
14
+ - en
15
+ library_name: transformers
16
+ ---
17
+
18
+ # ATLAS-Teach-8B-Instruct
19
+
20
+ An adaptive teaching model trained using the Reinforcement Collaborative Learning (RCL) framework. This is the supervised fine-tuning (SFT) checkpoint before reinforcement learning.
21
+
22
+ ## Model Details
23
+
24
+ - **Base Model**: Qwen/Qwen3-8B
25
+ - **Model Size**: 8B parameters
26
+ - **Training Stage**: Supervised Fine-tuning (Pre-RL)
27
+ - **Framework**: RCL (Reinforcement Collaborative Learning)
28
+
29
+ ## Training Data
30
+
31
+ Trained on `Arc-Intelligence/Arc-ATLAS-Teach-v0` dataset with RCL-specific formatting for adaptive teaching.
32
+
33
+ ## Intended Use
34
+
35
+ This model is designed for:
36
+ - Adaptive teaching based on student capability assessment
37
+ - Educational content generation
38
+ - Problem-solving assistance with tailored explanations
39
+
40
+ ## Training Configuration
41
+
42
+ - **Hardware**: 8x H100 GPUs
43
+ - **Framework**: RCL
44
+ - **Mixed Precision**: BF16
45
+
46
+ ## Adaptive Teaching Protocol
47
+
48
+ The model implements a two-pass teaching approach:
49
+
50
+ 1. **Diagnostic Probing**: Assesses student understanding with minimal interaction
51
+ 2. **Adaptive Teaching**: Generates tailored teaching based on diagnosed capability
52
+
53
+ ## Usage
54
+
55
+ ```python
56
+ from transformers import AutoModelForCausalLM, AutoTokenizer
57
+
58
+ model = AutoModelForCausalLM.from_pretrained("Arc-Intelligence/ATLAS-Teach-8B-Instruct")
59
+ tokenizer = AutoTokenizer.from_pretrained("Arc-Intelligence/ATLAS-Teach-8B-Instruct")
60
+
61
+ # Format your input according to the RCL teaching protocol
62
+ prompt = "Question: {your_question}\n\nProvide adaptive teaching:"
63
+ inputs = tokenizer(prompt, return_tensors="pt")
64
+ outputs = model.generate(**inputs)
65
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True)
66
+ ```
67
+
68
+ ## Limitations
69
+
70
+ - This is a pre-RL checkpoint; the full RCL training includes an additional RL phase
71
+ - Performance metrics on specific benchmarks are being evaluated
72
+
73
+ ## License
74
+
75
+ Apache 2.0