Testing of GGUF Llama3.2 3B
Browse files- README.md +1 -1
- app.py +2 -3
- models.lst +3 -0
README.md
CHANGED
|
@@ -5,7 +5,7 @@ colorFrom: green
|
|
| 5 |
colorTo: indigo
|
| 6 |
sdk: gradio
|
| 7 |
app_file: app.py
|
| 8 |
-
pinned:
|
| 9 |
license: cc-by-sa-4.0
|
| 10 |
short_description: SmallZOO runnigng SLMs directly on CPU with Llama.cpp&Python
|
| 11 |
---
|
|
|
|
| 5 |
colorTo: indigo
|
| 6 |
sdk: gradio
|
| 7 |
app_file: app.py
|
| 8 |
+
pinned: flase
|
| 9 |
license: cc-by-sa-4.0
|
| 10 |
short_description: SmallZOO runnigng SLMs directly on CPU with Llama.cpp&Python
|
| 11 |
---
|
app.py
CHANGED
|
@@ -2,10 +2,9 @@ import gradio as gr
|
|
| 2 |
from llama_cpp import Llama
|
| 3 |
import requests
|
| 4 |
|
| 5 |
-
|
| 6 |
llm = Llama.from_pretrained(
|
| 7 |
-
repo_id="
|
| 8 |
-
filename="*
|
| 9 |
verbose=True,
|
| 10 |
n_ctx=32768,
|
| 11 |
n_threads=2,
|
|
|
|
| 2 |
from llama_cpp import Llama
|
| 3 |
import requests
|
| 4 |
|
|
|
|
| 5 |
llm = Llama.from_pretrained(
|
| 6 |
+
repo_id="lmstudio-community/Llama-3.2-3B-Instruct-GGUF",
|
| 7 |
+
filename="*Q4_K_M.gguf",
|
| 8 |
verbose=True,
|
| 9 |
n_ctx=32768,
|
| 10 |
n_threads=2,
|
models.lst
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Stack of modesl to try:
|
| 2 |
+
|
| 3 |
+
https://huggingface.co/lmstudio-community/Llama-3.2-3B-Instruct-GGUF
|