Spaces:
Sleeping
Sleeping
remove unnecessary models
Browse files- gemmademo/_model.py +1 -14
gemmademo/_model.py
CHANGED
|
@@ -23,6 +23,7 @@ class LlamaCppGemmaModel:
|
|
| 23 |
_model_cache = {}
|
| 24 |
|
| 25 |
AVAILABLE_MODELS: Dict[str, Dict] = {
|
|
|
|
| 26 |
"gemma-3b": {
|
| 27 |
"model_path": "models/gemma-3-1b-it-Q5_K_M.gguf",
|
| 28 |
"repo_id": "bartowski/google_gemma-3-1b-it-GGUF", # Updated repo
|
|
@@ -31,13 +32,6 @@ class LlamaCppGemmaModel:
|
|
| 31 |
"type": "instruct",
|
| 32 |
},
|
| 33 |
"gemma-2b": {
|
| 34 |
-
"model_path": "models/gemma-2b.gguf",
|
| 35 |
-
"repo_id": "rahuldshetty/gemma-2b-gguf-quantized", # update to the actual repo id
|
| 36 |
-
"filename": "gemma-2b-Q4_K_M.gguf", # update to the actual filename
|
| 37 |
-
"description": "2B parameters, base model",
|
| 38 |
-
"type": "base",
|
| 39 |
-
},
|
| 40 |
-
"gemma-2b-it": {
|
| 41 |
"model_path": "models/gemma-2b-it.gguf",
|
| 42 |
"repo_id": "MaziyarPanahi/gemma-2b-it-GGUF", # update to the actual repo id
|
| 43 |
"filename": "gemma-2b-it.Q4_K_M.gguf", # update to the actual filename
|
|
@@ -45,13 +39,6 @@ class LlamaCppGemmaModel:
|
|
| 45 |
"type": "instruct",
|
| 46 |
},
|
| 47 |
"gemma-7b": {
|
| 48 |
-
"model_path": "models/gemma-7b.gguf",
|
| 49 |
-
"repo_id": "MaziyarPanahi/gemma-7b-GGUF", # update to the actual repo id
|
| 50 |
-
"filename": "gemma-7b.Q4_K_M.gguf", # update to the actual filename
|
| 51 |
-
"description": "7B parameters, instruction-tuned",
|
| 52 |
-
"type": "instruct",
|
| 53 |
-
},
|
| 54 |
-
"gemma-7b-it": {
|
| 55 |
"model_path": "models/gemma-7b-it.gguf",
|
| 56 |
"repo_id": "rahuldshetty/gemma-7b-it-gguf-quantized", # repository for the GGUF model
|
| 57 |
"filename": "gemma-7b-it-Q4_K_M.gguf", # updated filename for GGUF model
|
|
|
|
| 23 |
_model_cache = {}
|
| 24 |
|
| 25 |
AVAILABLE_MODELS: Dict[str, Dict] = {
|
| 26 |
+
# Does the job well.
|
| 27 |
"gemma-3b": {
|
| 28 |
"model_path": "models/gemma-3-1b-it-Q5_K_M.gguf",
|
| 29 |
"repo_id": "bartowski/google_gemma-3-1b-it-GGUF", # Updated repo
|
|
|
|
| 32 |
"type": "instruct",
|
| 33 |
},
|
| 34 |
"gemma-2b": {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
"model_path": "models/gemma-2b-it.gguf",
|
| 36 |
"repo_id": "MaziyarPanahi/gemma-2b-it-GGUF", # update to the actual repo id
|
| 37 |
"filename": "gemma-2b-it.Q4_K_M.gguf", # update to the actual filename
|
|
|
|
| 39 |
"type": "instruct",
|
| 40 |
},
|
| 41 |
"gemma-7b": {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
"model_path": "models/gemma-7b-it.gguf",
|
| 43 |
"repo_id": "rahuldshetty/gemma-7b-it-gguf-quantized", # repository for the GGUF model
|
| 44 |
"filename": "gemma-7b-it-Q4_K_M.gguf", # updated filename for GGUF model
|