Archi-medes commited on
Commit
e0e3cd3
·
verified ·
1 Parent(s): 70874b5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +119 -5
README.md CHANGED
@@ -1,5 +1,119 @@
1
- ---
2
- license: other
3
- license_name: lfm1.0
4
- license_link: https://huggingface.co/LiquidAI/LFM2-1.2B/blob/main/LICENSE
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: lfm1.0
4
+ license_link: https://huggingface.co/LiquidAI/LFM2-1.2B/blob/main/LICENSE
5
+ metrics:
6
+ - magic judge
7
+ base_model:
8
+ - LiquidAI/LFM2-1.2B
9
+ tags:
10
+ - lmstudio
11
+ - madlabOSS
12
+ - magic judge
13
+ ---
14
+
15
+ # LMS Guide 350m
16
+
17
+ ## 🧠 Overview
18
+ The **LMS Guide 350m** is part of the **MadlabOSS LM Studio Guide** family — a lineup of small, efficient, and highly aligned assistant models trained specifically to provide deterministic, hallucination‑resistant guidance for LM Studio users.
19
+
20
+ This model is trained on a curated dataset of LM Studio–specific instructions, workflows, troubleshooting steps, and conceptual explanations.
21
+
22
+ ---
23
+
24
+ ## 🚀 Intended Use
25
+ This model is optimized for:
26
+
27
+ - LM Studio onboarding
28
+ - workflow explanations
29
+ - feature descriptions
30
+ - troubleshooting guidance
31
+ - plugin/server integration help
32
+ - safe, deterministic assistant behavior
33
+
34
+ It is **not** intended as a general‑purpose chatbot.
35
+
36
+ ---
37
+
38
+ ## 🧩 Model Details
39
+
40
+ **Base Model:** LFM2‑1.2B
41
+
42
+ **Parameter Count:** 1.2 Billion
43
+
44
+ **Training Type:** Supervised fine‑tuning
45
+
46
+ **Sequence Length:** 1024
47
+
48
+ **Precision:** FP16
49
+
50
+ **Framework:** PyTorch / Transformers
51
+
52
+ ---
53
+
54
+ ## 📦 Training Data
55
+ The model was trained on:
56
+
57
+ - **6,000+ LM Studio–specific instruction/response pairs**
58
+ - Clean, domain‑specific, ontology‑consistent data
59
+ - Minor general‑purpose conversational data
60
+ - No web‑scraped content
61
+ - Full LM Studio Documentation
62
+
63
+ A 36k+ expanded dataset is planned for v2.0.
64
+
65
+ ---
66
+
67
+ ## 🏋️ Training Procedure
68
+
69
+ ### **Hyperparameters**
70
+ - Epochs: 6
71
+ - Batch size: 16
72
+ - Learning rate: cosine schedule, peak ~4e‑5
73
+ - Optimizer: AdamW
74
+ - Gradient clipping: 1.0
75
+ - Gradient accumulation: 1
76
+
77
+ ### **Hardware**
78
+ Training was performed on:
79
+
80
+ - RTX 6000 Ada (96GB) (1.2b + 2.6b)
81
+ - Dual RTX 3090 (Magic Judge)
82
+ - RTX 3070 (for 0.35B + 0.7b)
83
+
84
+ ---
85
+
86
+ ## 📊 Evaluation
87
+
88
+ ### **Judge Score**
89
+ Semantic correctness, ontology adherence, and hallucination resistance.
90
+
91
+
92
+ ### **Qualitative Behavior**
93
+ - Strong adherence to LM Studio terminology
94
+ - Low hallucination rate
95
+ - Deterministic, predictable responses
96
+ - Not optimized for open‑domain reasoning
97
+
98
+ ---
99
+
100
+ ## 🔒 Safety
101
+ This model is trained exclusively on LM Studio–specific content.
102
+ It avoids hallucinating non‑existent LM Studio features and adheres to a strict ontology.
103
+
104
+ It is **not** designed for:
105
+
106
+ - political content
107
+ - medical advice
108
+ - legal advice
109
+ - general‑purpose conversation
110
+
111
+ ---
112
+
113
+ ## ⚠️ Limitations
114
+ - Not a general assistant
115
+ - Not trained for coding, math, or open‑domain reasoning
116
+ - May refuse tasks outside LM Studio scope
117
+ - Static accuracy metrics underestimate real performance
118
+
119
+ ---