Update README.md
Browse files
README.md
CHANGED
|
@@ -145,6 +145,20 @@ print(run_command("rm test.txt")) # (empty)
|
|
| 145 |
print(run_command("ls")) # backup.txt
|
| 146 |
```
|
| 147 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 148 |
## Supported Commands
|
| 149 |
|
| 150 |
| Command | Description | Example |
|
|
|
|
| 145 |
print(run_command("ls")) # backup.txt
|
| 146 |
```
|
| 147 |
|
| 148 |
+
## Quantized Versions
|
| 149 |
+
|
| 150 |
+
GGUF quantizations are available for CPU inference and lower memory usage:
|
| 151 |
+
|
| 152 |
+
**[LaaLM-exp-v1-GGUF](https://huggingface.co/ereniko/LaaLM-exp-v1-GGUF)**
|
| 153 |
+
|
| 154 |
+
Includes Q2_K through fp16 quantizations (1.27GB - 6.18GB) for use with:
|
| 155 |
+
- llama.cpp
|
| 156 |
+
- Ollama
|
| 157 |
+
- llama-cpp-python
|
| 158 |
+
- Other GGUF-compatible tools
|
| 159 |
+
|
| 160 |
+
Recommended: Q4_K_M (1.93GB) for best quality/size balance.
|
| 161 |
+
|
| 162 |
## Supported Commands
|
| 163 |
|
| 164 |
| Command | Description | Example |
|