ereniko commited on
Commit
bb90bd8
·
verified ·
1 Parent(s): cf80e2b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -0
README.md CHANGED
@@ -145,6 +145,20 @@ print(run_command("rm test.txt")) # (empty)
145
  print(run_command("ls")) # backup.txt
146
  ```
147
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
148
  ## Supported Commands
149
 
150
  | Command | Description | Example |
 
145
  print(run_command("ls")) # backup.txt
146
  ```
147
 
148
+ ## Quantized Versions
149
+
150
+ GGUF quantizations are available for CPU inference and lower memory usage:
151
+
152
+ **[LaaLM-exp-v1-GGUF](https://huggingface.co/ereniko/LaaLM-exp-v1-GGUF)**
153
+
154
+ Includes Q2_K through fp16 quantizations (1.27GB - 6.18GB) for use with:
155
+ - llama.cpp
156
+ - Ollama
157
+ - llama-cpp-python
158
+ - Other GGUF-compatible tools
159
+
160
+ Recommended: Q4_K_M (1.93GB) for best quality/size balance.
161
+
162
  ## Supported Commands
163
 
164
  | Command | Description | Example |