Update README.md
Browse files
README.md
CHANGED
|
@@ -33,12 +33,12 @@ This model was made with the Micro Distillery app available at:
|
|
| 33 |
|
| 34 |
webxos.netlify.app/MICROD
|
| 35 |
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
|
| 43 |
<div id="app">
|
| 44 |
<!-- TOP BAR -->
|
|
@@ -94,7 +94,7 @@ If you use this data in research, please cite:
|
|
| 94 |
}
|
| 95 |
|
| 96 |
|
| 97 |
-
### Using Transformers
|
| 98 |
```python
|
| 99 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 100 |
|
|
|
|
| 33 |
|
| 34 |
webxos.netlify.app/MICROD
|
| 35 |
|
| 36 |
+
-Model Distillation Training: Simulate GRPO optimization with VAE filtering for small LLMs (42M-345M params).
|
| 37 |
+
-Policy Experimentation: Test group sizes, KL penalties, cache reuse for RLHF-like training.
|
| 38 |
+
-VAE Filtering: Apply latent space compression to improve distillation quality.
|
| 39 |
+
-Sandbox Testing: Execute safe Python code with feedback masking.
|
| 40 |
+
-Export & Deployment: Generate deployable models for inference in various frameworks.
|
| 41 |
+
-Offline Usage: PWA supports offline training simulation and exports.
|
| 42 |
|
| 43 |
<div id="app">
|
| 44 |
<!-- TOP BAR -->
|
|
|
|
| 94 |
}
|
| 95 |
|
| 96 |
|
| 97 |
+
### EXAMPLE: Using Transformers
|
| 98 |
```python
|
| 99 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 100 |
|