Archi-medes commited on
Commit
8fb2acb
·
verified ·
1 Parent(s): 63392c3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +115 -5
README.md CHANGED
@@ -1,5 +1,115 @@
1
- ---
2
- license: other
3
- license_name: lfm1.0
4
- license_link: https://huggingface.co/LiquidAI/LFM2-350M/blob/main/LICENSE
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: lfm1.0
4
+ license_link: https://huggingface.co/LiquidAI/LFM2-350M/blob/main/LICENSE
5
+ metrics:
6
+ - magic judge
7
+ base_model:
8
+ - LiquidAI/LFM2-350M
9
+ tags:
10
+ - lm-studio
11
+ - madlabOSS
12
+ ---
13
+
14
+ # LMS Guide 350m
15
+
16
+ ## 🧠 Overview
17
+ The **LMS Guide 350m** is part of the **MadlabOSS LM Studio Guide** family — a lineup of small, efficient, and highly aligned assistant models trained specifically to provide deterministic, hallucination‑resistant guidance for LM Studio users.
18
+
19
+ This model is trained on a curated dataset of LM Studio–specific instructions, workflows, troubleshooting steps, and conceptual explanations.
20
+
21
+ ---
22
+
23
+ ## 🚀 Intended Use
24
+ This model is optimized for:
25
+
26
+ - LM Studio onboarding
27
+ - workflow explanations
28
+ - feature descriptions
29
+ - troubleshooting guidance
30
+ - plugin/server integration help
31
+ - safe, deterministic assistant behavior
32
+
33
+ It is **not** intended as a general‑purpose chatbot.
34
+
35
+ ---
36
+
37
+ ## 🧩 Model Details
38
+
39
+ **Base Model:** LFM2‑350m
40
+ **Parameter Count:** 350 Million
41
+ **Training Type:** Supervised fine‑tuning
42
+ **Sequence Length:** 320
43
+ **Precision:** FP16
44
+ **Framework:** PyTorch / Transformers
45
+
46
+ ---
47
+
48
+ ## 📦 Training Data
49
+ The model was trained on:
50
+
51
+ - **6,000+ LM Studio–specific instruction/response pairs**
52
+ - Clean, domain‑specific, ontology‑consistent data
53
+ - Minor general‑purpose conversational data
54
+ - No web‑scraped content
55
+ - Full LM Studio Documentation
56
+
57
+ A 36k+ expanded dataset is planned for v2.0.
58
+
59
+ ---
60
+
61
+ ## 🏋️ Training Procedure
62
+
63
+ ### **Hyperparameters**
64
+ - Epochs: 6
65
+ - Batch size:
66
+ 0.35b + 0.7b = 4
67
+ 1.2b + 2.6b = 16
68
+ - Learning rate: cosine schedule, peak ~4e‑5
69
+ - Optimizer: AdamW
70
+ - Gradient clipping: 1.0
71
+ - Gradient accumulation: 1
72
+
73
+ ### **Hardware**
74
+ Training was performed on:
75
+
76
+ - RTX 6000 Ada (96GB) (1.2b + 2.6b)
77
+ - Dual RTX 3090 (Magic Judge)
78
+ - RTX 3070 (for 0.35B + 0.7b)
79
+
80
+ ---
81
+
82
+ ## 📊 Evaluation
83
+
84
+ ### **Judge Score**
85
+ Semantic correctness, ontology adherence, and hallucination resistance.
86
+
87
+
88
+ ### **Qualitative Behavior**
89
+ - Strong adherence to LM Studio terminology
90
+ - Low hallucination rate
91
+ - Deterministic, predictable responses
92
+ - Not optimized for open‑domain reasoning
93
+
94
+ ---
95
+
96
+ ## 🔒 Safety
97
+ This model is trained exclusively on LM Studio–specific content.
98
+ It avoids hallucinating non‑existent LM Studio features and adheres to a strict ontology.
99
+
100
+ It is **not** designed for:
101
+
102
+ - political content
103
+ - medical advice
104
+ - legal advice
105
+ - general‑purpose conversation
106
+
107
+ ---
108
+
109
+ ## ⚠️ Limitations
110
+ - Not a general assistant
111
+ - Not trained for coding, math, or open‑domain reasoning
112
+ - May refuse tasks outside LM Studio scope
113
+ - Static accuracy metrics underestimate real performance
114
+
115
+ ---