File size: 3,390 Bytes
8028e3b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
c8a56b4
8028e3b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f5bf31f
8028e3b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
---
license: other
license_name: lfm1.0
license_link: https://huggingface.co/LiquidAI/LFM2-2.6B/blob/main/LICENSE
metrics:
- Synthetic Data Expansion Benchmark
base_model:
- LiquidAI/LFM2-2.6B
tags:
- lmstudio
- madlabOSS
- synthetic data generator
---

# Madlab Synthetic Data Generator

## 🧠 Overview
The **Madlab SDG 2.6B** is part of the **MadlabOSS Synthetic Data Generator** family — a suite of small, efficient synthetic data generators designed for rule‑consistent, semantically coherent variation.  
This model was trained on a closed-source dataset created through a multi-stage synthetic data generation process using a modified Madlab training pipeline.

---

## 🚀 Intended Use
This model is optimized for:

- Madlab synthetic data generation

It is **not** intended as a general-purpose chatbot.

---

## 🧩 Model Details

**Base Model:** LFM2-2.6B  
**Parameter Count:** 2.6 Billion  
**Training Type:** Supervised fine-tuning  
**Sequence Length:** 1024  
**Precision:** FP16  
**Framework:** PyTorch / Transformers  

---

## 📦 Training Data
The model was trained on:

- **1444 compressed and encoded dataset pairs**  
- High variation in output  
- Preservation of semantic meaning  
- Data entirely generated with Madlab  

---

## 🏋️ Training Procedure

### **Hyperparameters**
- Epochs: 6  
- Batch size: 48  
- Learning rate: cosine schedule, peak ~4e-5  
- Optimizer: AdamW  
- Gradient clipping: 1.0  
- Gradient accumulation: 1  

### **Hardware**
Training was performed on:

- RTX 6000 Blackwell (96GB)

---

## 📊 Evaluation

### **Synthetic Data Expansion Benchmark**
A curated set of 30 input/target pairs was programmatically expanded using a Python script.  
Metrics include seed pairs covered, total variation count, and semantic quality.  
The task is to generate 5 variations of each incoming pair.

| Run | Model | Semantic Quality | Variations | Seeds Covered | Efficiency (Variations/Param) | Dataset |
|-----|------------|---------------|------------|---------------|-------------------------------|--------------|
| 1 | LFM2-350M-16 | 6.5 | 94 | 23 | 268.57 | Madlab sdg small |
| 2 | LFM2-350M-16 | 3.5 | 46 | 11 | 131.43 | base model |
| 3 | LFM2-350M-f16 | 6.5 | 97 | 22 | 277.14 | Madlab sdg small |
| 4 | Qwen3-coder-30B-instruct-q8 | 8.2 | 149 | 26 | 4.97 | base model |
| 5 | LFM2-350M-f16 | 7.5 | 136 | 21 | 388.57 | Madlab sdg medium |
| 6 | LFM2-2.6B-f16 | 9.0 | 137 | 25 | 52.69 | Madlab sdg medium |
| 7 | LFM2-2.6B-f16 | 9.9 | 180 | 25 | 69.23 | Madlab sdg large |
| 8 | LFM2-2.6B-f16 | 6.2 | 157 | 20 | 60.38 | Madlab sdg test |
| 9 | LFM2-2.6B-f16 | 10.0 | 248 | 27 | 95.38 | Madlab sdg large |
| 10 | Qwen3-235B-q3-k_m | 9.5 | 150 | 27 | 0.64 | base model |
| 11 | LFM2.5-1.2B-instruct-f16 | 9.1 | 244 | 30 | 203.33 | Madlab sdg large |

### **Qualitative Behavior**
- Overperforms in variation count  
- Maintains strict semantic correctness  

---

## 🔒 Safety
This model is a synthetic data generator. It is not designed for conversational use and is not suitable for anything other than generating synthetic datasets.

It is **not** designed for:

- Political advice  
- Medical advice  
- Legal advice  
- General-purpose conversation  

---

## ⚠️ Limitations
- Not a general assistant  
- Not trained for coding, math, or open-domain reasoning  
- May refuse tasks outside the Madlab SDG scope  

---