Spaces:
Sleeping
Sleeping
remove unnecessary model
Browse files- gemmademo/_model.py +1 -8
gemmademo/_model.py
CHANGED
|
@@ -30,13 +30,6 @@ class LlamaCppGemmaModel:
|
|
| 30 |
"description": "3B parameters, instruction-tuned (Q5_K_M)",
|
| 31 |
"type": "instruct",
|
| 32 |
},
|
| 33 |
-
"gemma-3b-q6": {
|
| 34 |
-
"model_path": "models/gemma-3-1b-it-Q6_K.gguf",
|
| 35 |
-
"repo_id": "bartowski/google_gemma-3-1b-it-GGUF", # Updated repo
|
| 36 |
-
"filename": "google_gemma-3-1b-it-Q6_K.gguf", # Higher quality quantization
|
| 37 |
-
"description": "3B parameters, instruction-tuned (Q6_K)",
|
| 38 |
-
"type": "instruct",
|
| 39 |
-
},
|
| 40 |
"gemma-2b": {
|
| 41 |
"model_path": "models/gemma-2b.gguf",
|
| 42 |
"repo_id": "rahuldshetty/gemma-2b-gguf-quantized", # update to the actual repo id
|
|
@@ -58,7 +51,7 @@ class LlamaCppGemmaModel:
|
|
| 58 |
"description": "7B parameters, instruction-tuned",
|
| 59 |
"type": "instruct",
|
| 60 |
},
|
| 61 |
-
"gemma-7b
|
| 62 |
"model_path": "models/gemma-7b.gguf",
|
| 63 |
"repo_id": "rahuldshetty/gemma-7b-it-gguf-quantized", # repository for the GGUF model
|
| 64 |
"filename": "gemma-7b-it-Q4_K_M.gguf", # updated filename for GGUF model
|
|
|
|
| 30 |
"description": "3B parameters, instruction-tuned (Q5_K_M)",
|
| 31 |
"type": "instruct",
|
| 32 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
"gemma-2b": {
|
| 34 |
"model_path": "models/gemma-2b.gguf",
|
| 35 |
"repo_id": "rahuldshetty/gemma-2b-gguf-quantized", # update to the actual repo id
|
|
|
|
| 51 |
"description": "7B parameters, instruction-tuned",
|
| 52 |
"type": "instruct",
|
| 53 |
},
|
| 54 |
+
"gemma-7b": {
|
| 55 |
"model_path": "models/gemma-7b.gguf",
|
| 56 |
"repo_id": "rahuldshetty/gemma-7b-it-gguf-quantized", # repository for the GGUF model
|
| 57 |
"filename": "gemma-7b-it-Q4_K_M.gguf", # updated filename for GGUF model
|