webxos commited on
Commit
25c5968
·
verified ·
1 Parent(s): 168dfa1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -33,12 +33,12 @@ This model was made with the Micro Distillery app available at:
33
 
34
  webxos.netlify.app/MICROD
35
 
36
- -Model Distillation Training: Simulate GRPO optimization with VAE filtering for small LLMs (42M-345M params).
37
- -Policy Experimentation: Test group sizes, KL penalties, cache reuse for RLHF-like training.
38
- -VAE Filtering: Apply latent space compression to improve distillation quality.
39
- -Sandbox Testing: Execute safe Python code with feedback masking.
40
- -Export & Deployment: Generate deployable models for inference in various frameworks.
41
- -Offline Usage: PWA supports offline training simulation and exports.
42
 
43
  <div id="app">
44
  <!-- TOP BAR -->
@@ -94,7 +94,7 @@ If you use this data in research, please cite:
94
  }
95
 
96
 
97
- ### Using Transformers
98
  ```python
99
  from transformers import AutoModelForCausalLM, AutoTokenizer
100
 
 
33
 
34
  webxos.netlify.app/MICROD
35
 
36
+ -Model Distillation Training: Simulate GRPO optimization with VAE filtering for small LLMs (42M-345M params).
37
+ -Policy Experimentation: Test group sizes, KL penalties, cache reuse for RLHF-like training.
38
+ -VAE Filtering: Apply latent space compression to improve distillation quality.
39
+ -Sandbox Testing: Execute safe Python code with feedback masking.
40
+ -Export & Deployment: Generate deployable models for inference in various frameworks.
41
+ -Offline Usage: PWA supports offline training simulation and exports.
42
 
43
  <div id="app">
44
  <!-- TOP BAR -->
 
94
  }
95
 
96
 
97
+ ### EXAMPLE: Using Transformers
98
  ```python
99
  from transformers import AutoModelForCausalLM, AutoTokenizer
100