Update README.md
Browse files
README.md
CHANGED
|
@@ -2,6 +2,9 @@
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- google-t5/t5-base
|
|
|
|
|
|
|
|
|
|
| 5 |
---
|
| 6 |
|
| 7 |
# LaaLM — Linux as a Language Model
|
|
@@ -117,5 +120,4 @@ print(run_command("Current directory: /home/user", "pwd"))
|
|
| 117 |
|
| 118 |
Very big thanks to [@mradermacher](https://huggingface.co/mradermacher) for spending time and compute to generate GGUF Quantized versions of LaaLM-v1!
|
| 119 |
|
| 120 |
-
Check out the quantized versions he made [here](https://huggingface.co/mradermacher/LaaLM-v1-GGUF). With these you can run LaaLM-v1 on low tier GPUs or CPUs!
|
| 121 |
-
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
base_model:
|
| 4 |
- google-t5/t5-base
|
| 5 |
+
language:
|
| 6 |
+
- en
|
| 7 |
+
library_name: transformers
|
| 8 |
---
|
| 9 |
|
| 10 |
# LaaLM — Linux as a Language Model
|
|
|
|
| 120 |
|
| 121 |
Very big thanks to [@mradermacher](https://huggingface.co/mradermacher) for spending time and compute to generate GGUF Quantized versions of LaaLM-v1!
|
| 122 |
|
| 123 |
+
Check out the quantized versions he made [here](https://huggingface.co/mradermacher/LaaLM-v1-GGUF). With these you can run LaaLM-v1 on low tier GPUs or CPUs!
|
|
|