Bootstrap recommended models dataset with 39 imported entries
Browse filesImport the original bundled catalog into per-model metadata entries.
- README.md +8 -0
- models/DeepSeek-R1-Distill-70B-Q4_K_M/metadata.json +12 -0
- models/Llama-3.1-405B-Instruct-Q2_K/metadata.json +12 -0
- models/Llama-3.3-70B-Instruct-Q4_K_M/metadata.json +12 -0
- models/Mixtral-8x22B-Instruct-Q4_K_M/metadata.json +16 -0
- models/Qwen%2FQwen2.5-0.5B-Instruct-GGUF%2Fqwen2.5-0.5b-instruct-q4_k_m.gguf/metadata.json +12 -0
- models/Qwen%2FQwen2.5-3B-Instruct-GGUF%2Fqwen2.5-3b-instruct-q4_k_m.gguf/metadata.json +12 -0
- models/Qwen%2FQwen2.5-Coder-32B-Instruct-GGUF%2Fqwen2.5-coder-32b-instruct-q4_k_m.gguf/metadata.json +12 -0
- models/Qwen%2FQwen2.5-Coder-7B-Instruct-GGUF%2Fqwen2.5-coder-7b-instruct-q4_k_m.gguf/metadata.json +12 -0
- models/Qwen%2FQwen3-Coder-Next-GGUF%3AQwen3-Coder-Next-Q4_K_M/metadata.json +25 -0
- models/Qwen2.5-72B-Instruct-Q4_K_M/metadata.json +12 -0
- models/Qwen3-235B-A22B-Q4_K_M/metadata.json +16 -0
- models/Qwen3.5-27B-Q4_K_M/metadata.json +15 -0
- models/bartowski%2FHermes-2-Pro-Mistral-7B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2FLlama-3.2-1B-Instruct-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2FLlama-3.2-3B-Instruct-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2FQwen2.5-14B-Instruct-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2FQwen2.5-32B-Instruct-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2FQwen2.5-Coder-14B-Instruct-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2Fgoogle_gemma-3-1b-it-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/bartowski%2Fgoogle_gemma-3-27b-it-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/glogwa68%2FLlama-4-scout-GGUF%3AQ4_K_M/metadata.json +16 -0
- models/unsloth%2FDeepSeek-R1-Distill-Qwen-14B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FDeepSeek-R1-Distill-Qwen-32B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FDevstral-Small-2505-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FGLM-4-32B-0414-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FGLM-4.7-Flash-GGUF%3AQ4_K_M/metadata.json +16 -0
- models/unsloth%2FMiniMax-M2.5-GGUF%3AQ4_K_M/metadata.json +29 -0
- models/unsloth%2FMistral-Small-3.1-24B-Instruct-2503-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FQwen3-0.6B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FQwen3-14B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FQwen3-30B-A3B-GGUF%3AQ4_K_M/metadata.json +16 -0
- models/unsloth%2FQwen3-32B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FQwen3-4B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FQwen3-8B-GGUF%3AQ4_K_M/metadata.json +12 -0
- models/unsloth%2FQwen3-Coder-30B-A3B-Instruct-GGUF%3AQ4_K_M/metadata.json +16 -0
- models/unsloth%2FQwen3.5-0.8B-GGUF%3AQ4_K_M/metadata.json +15 -0
- models/unsloth%2FQwen3.5-4B-GGUF%3AQ4_K_M/metadata.json +15 -0
- models/unsloth%2FQwen3.5-9B-GGUF%3AQ4_K_M/metadata.json +15 -0
- models/unsloth%2Fgemma-3-12b-it-GGUF%3AQ4_K_M/metadata.json +12 -0
README.md
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Recommended Models
|
| 2 |
+
|
| 3 |
+
This dataset stores curated recommended model metadata for `mesh-llm`.
|
| 4 |
+
|
| 5 |
+
Layout:
|
| 6 |
+
- `models/<urlencoded-model-id>/metadata.json`
|
| 7 |
+
|
| 8 |
+
Each entry stores the curated model `id`, display `name`, resolved download `url`, sizing, description, draft hint, MoE hints, split assets, and optional `mmproj`.
|
models/DeepSeek-R1-Distill-70B-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "DeepSeek-R1-Distill-70B-Q4_K_M",
|
| 3 |
+
"file": "DeepSeek-R1-Distill-70B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/deepseek-r1/blobs/sha256:4cd576d9aa16961244012223abf01445567b061f1814b57dfef699e4cf8df339",
|
| 5 |
+
"size": "43GB",
|
| 6 |
+
"description": "DeepSeek R1 distilled to 70B (Qwen2.5-based), strong reasoning",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "DeepSeek-R1-Distill-70B-Q4_K_M"
|
| 12 |
+
}
|
models/Llama-3.1-405B-Instruct-Q2_K/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Llama-3.1-405B-Instruct-Q2_K",
|
| 3 |
+
"file": "Llama-3.1-405B-Instruct-Q2_K.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/llama3.1/blobs/sha256:e7e1972e5b13caead8a8dd9c94f4a0dec59ac2d9dd52e0cd1c067e6077eb4677",
|
| 5 |
+
"size": "149GB",
|
| 6 |
+
"description": "Llama 3.1 405B Instruct Q2_K, largest dense model",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Llama-3.1-405B-Instruct-Q2_K"
|
| 12 |
+
}
|
models/Llama-3.3-70B-Instruct-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Llama-3.3-70B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Llama-3.3-70B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/llama3.3/blobs/sha256:4824460d29f2058aaf6e1118a63a7a197a09bed509f0e7d4e2efb1ee273b447d",
|
| 5 |
+
"size": "43GB",
|
| 6 |
+
"description": "Meta Llama 3.3 70B, strong all-around",
|
| 7 |
+
"draft": "Llama-3.2-1B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Llama-3.3-70B-Instruct-Q4_K_M"
|
| 12 |
+
}
|
models/Mixtral-8x22B-Instruct-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Mixtral-8x22B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Mixtral-8x22B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/mixtral/blobs/sha256:f3329ad0c787f4f73cab99e8c877bb76403060561dd0caa318127683c87bbcb4",
|
| 5 |
+
"size": "86GB",
|
| 6 |
+
"description": "Mixtral 8x22B MoE, 8 experts top-2",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 8,
|
| 10 |
+
"n_expert_used": 2,
|
| 11 |
+
"min_experts_per_node": 4
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [],
|
| 14 |
+
"mmproj": null,
|
| 15 |
+
"id": "Mixtral-8x22B-Instruct-Q4_K_M"
|
| 16 |
+
}
|
models/Qwen%2FQwen2.5-0.5B-Instruct-GGUF%2Fqwen2.5-0.5b-instruct-q4_k_m.gguf/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-0.5B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/qwen2.5-0.5b-instruct-q4_k_m.gguf",
|
| 5 |
+
"size": "491MB",
|
| 6 |
+
"description": "Draft for Qwen2.5 and DeepSeek-R1-Distill models",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF/qwen2.5-0.5b-instruct-q4_k_m.gguf"
|
| 12 |
+
}
|
models/Qwen%2FQwen2.5-3B-Instruct-GGUF%2Fqwen2.5-3b-instruct-q4_k_m.gguf/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-3B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-3B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct-GGUF/resolve/main/qwen2.5-3b-instruct-q4_k_m.gguf",
|
| 5 |
+
"size": "2.1GB",
|
| 6 |
+
"description": "Small & fast general chat",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Qwen/Qwen2.5-3B-Instruct-GGUF/qwen2.5-3b-instruct-q4_k_m.gguf"
|
| 12 |
+
}
|
models/Qwen%2FQwen2.5-Coder-32B-Instruct-GGUF%2Fqwen2.5-coder-32b-instruct-q4_k_m.gguf/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-Coder-32B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-Coder-32B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct-GGUF/resolve/main/qwen2.5-coder-32b-instruct-q4_k_m.gguf",
|
| 5 |
+
"size": "20GB",
|
| 6 |
+
"description": "Top-tier code gen, matches GPT-4o on code",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Qwen/Qwen2.5-Coder-32B-Instruct-GGUF/qwen2.5-coder-32b-instruct-q4_k_m.gguf"
|
| 12 |
+
}
|
models/Qwen%2FQwen2.5-Coder-7B-Instruct-GGUF%2Fqwen2.5-coder-7b-instruct-q4_k_m.gguf/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-Coder-7B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-Coder-7B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct-GGUF/resolve/main/qwen2.5-coder-7b-instruct-q4_k_m.gguf",
|
| 5 |
+
"size": "4.4GB",
|
| 6 |
+
"description": "Code generation & completion",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Qwen/Qwen2.5-Coder-7B-Instruct-GGUF/qwen2.5-coder-7b-instruct-q4_k_m.gguf"
|
| 12 |
+
}
|
models/Qwen%2FQwen3-Coder-Next-GGUF%3AQwen3-Coder-Next-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-Coder-Next-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-Coder-Next-Q4_K_M-00001-of-00004.gguf",
|
| 4 |
+
"url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00001-of-00004.gguf",
|
| 5 |
+
"size": "48GB",
|
| 6 |
+
"description": "Qwen3 Coder Next ~85B dense, frontier coding model",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [
|
| 10 |
+
{
|
| 11 |
+
"file": "Qwen3-Coder-Next-Q4_K_M-00002-of-00004.gguf",
|
| 12 |
+
"url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00002-of-00004.gguf"
|
| 13 |
+
},
|
| 14 |
+
{
|
| 15 |
+
"file": "Qwen3-Coder-Next-Q4_K_M-00003-of-00004.gguf",
|
| 16 |
+
"url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00003-of-00004.gguf"
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"file": "Qwen3-Coder-Next-Q4_K_M-00004-of-00004.gguf",
|
| 20 |
+
"url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00004-of-00004.gguf"
|
| 21 |
+
}
|
| 22 |
+
],
|
| 23 |
+
"mmproj": null,
|
| 24 |
+
"id": "Qwen/Qwen3-Coder-Next-GGUF:Qwen3-Coder-Next-Q4_K_M"
|
| 25 |
+
}
|
models/Qwen2.5-72B-Instruct-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-72B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-72B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/qwen2.5/blobs/sha256:6e7fdda508e91cb0f63de5c15ff79ac63a1584ccafd751c07ca12b7f442101b8",
|
| 5 |
+
"size": "47GB",
|
| 6 |
+
"description": "Flagship Qwen2.5, great tensor split showcase",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "Qwen2.5-72B-Instruct-Q4_K_M"
|
| 12 |
+
}
|
models/Qwen3-235B-A22B-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-235B-A22B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-235B-A22B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/qwen3/blobs/sha256:aeacdadecbed8a07e42026d1a1d3cd30715bb2994ebe4e4ca4009e1a4abe8d5d",
|
| 5 |
+
"size": "142GB",
|
| 6 |
+
"description": "Qwen3 235B MoE A22B active, 128 experts top-8",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 128,
|
| 10 |
+
"n_expert_used": 8,
|
| 11 |
+
"min_experts_per_node": 46
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [],
|
| 14 |
+
"mmproj": null,
|
| 15 |
+
"id": "Qwen3-235B-A22B-Q4_K_M"
|
| 16 |
+
}
|
models/Qwen3.5-27B-Q4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3.5-27B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3.5-27B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://registry.ollama.ai/v2/library/qwen3.5/blobs/sha256:d4b8b4f4c350f5d322dc8235175eeae02d32c6f3fd70bdb9ea481e3abb7d7fc4",
|
| 5 |
+
"size": "17GB",
|
| 6 |
+
"description": "Qwen3.5 27B, vision + text, strong reasoning and coding",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": {
|
| 11 |
+
"file": "Qwen3.5-27B-mmproj-BF16.gguf",
|
| 12 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-27B-GGUF/resolve/main/mmproj-BF16.gguf"
|
| 13 |
+
},
|
| 14 |
+
"id": "Qwen3.5-27B-Q4_K_M"
|
| 15 |
+
}
|
models/bartowski%2FHermes-2-Pro-Mistral-7B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Hermes-2-Pro-Mistral-7B-Q4_K_M",
|
| 3 |
+
"file": "Hermes-2-Pro-Mistral-7B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/Hermes-2-Pro-Mistral-7B-GGUF/resolve/main/Hermes-2-Pro-Mistral-7B-Q4_K_M.gguf",
|
| 5 |
+
"size": "4.4GB",
|
| 6 |
+
"description": "Goose default, strong tool calling for agents",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/Hermes-2-Pro-Mistral-7B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2FLlama-3.2-1B-Instruct-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Llama-3.2-1B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Llama-3.2-1B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF/resolve/main/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
|
| 5 |
+
"size": "760MB",
|
| 6 |
+
"description": "Draft for Llama 3.x and Llama 4 models",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/Llama-3.2-1B-Instruct-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2FLlama-3.2-3B-Instruct-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Llama-3.2-3B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Llama-3.2-3B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_M.gguf",
|
| 5 |
+
"size": "2.0GB",
|
| 6 |
+
"description": "Meta Llama 3.2, goose default, good tool calling",
|
| 7 |
+
"draft": "Llama-3.2-1B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2FQwen2.5-14B-Instruct-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-14B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-14B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/Qwen2.5-14B-Instruct-GGUF/resolve/main/Qwen2.5-14B-Instruct-Q4_K_M.gguf",
|
| 5 |
+
"size": "9.0GB",
|
| 6 |
+
"description": "Solid general chat",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/Qwen2.5-14B-Instruct-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2FQwen2.5-32B-Instruct-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-32B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-32B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/Qwen2.5-32B-Instruct-GGUF/resolve/main/Qwen2.5-32B-Instruct-Q4_K_M.gguf",
|
| 5 |
+
"size": "20GB",
|
| 6 |
+
"description": "Proven general chat",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/Qwen2.5-32B-Instruct-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2FQwen2.5-Coder-14B-Instruct-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen2.5-Coder-14B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen2.5-Coder-14B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/Qwen2.5-Coder-14B-Instruct-GGUF/resolve/main/Qwen2.5-Coder-14B-Instruct-Q4_K_M.gguf",
|
| 5 |
+
"size": "9.0GB",
|
| 6 |
+
"description": "Strong code gen, fills gap between 7B and 32B",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/Qwen2.5-Coder-14B-Instruct-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2Fgoogle_gemma-3-1b-it-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Gemma-3-1B-it-Q4_K_M",
|
| 3 |
+
"file": "Gemma-3-1B-it-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/google_gemma-3-1b-it-GGUF/resolve/main/google_gemma-3-1b-it-Q4_K_M.gguf",
|
| 5 |
+
"size": "780MB",
|
| 6 |
+
"description": "Draft for Gemma 3 models",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/google_gemma-3-1b-it-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/bartowski%2Fgoogle_gemma-3-27b-it-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Gemma-3-27B-it-Q4_K_M",
|
| 3 |
+
"file": "Gemma-3-27B-it-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/bartowski/google_gemma-3-27b-it-GGUF/resolve/main/google_gemma-3-27b-it-Q4_K_M.gguf",
|
| 5 |
+
"size": "17GB",
|
| 6 |
+
"description": "Google Gemma 3 27B, strong reasoning",
|
| 7 |
+
"draft": "Gemma-3-1B-it-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "bartowski/google_gemma-3-27b-it-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/glogwa68%2FLlama-4-scout-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Llama-4-Scout-Q4_K_M",
|
| 3 |
+
"file": "Llama-4-Scout-4bit-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/glogwa68/Llama-4-scout-GGUF/resolve/main/Llama-4-Scout-4bit-Q4_K_M.gguf",
|
| 5 |
+
"size": "22.5GB",
|
| 6 |
+
"description": "MoE 109B/17B active, 16 experts top-1, Meta latest, tool calling",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 16,
|
| 10 |
+
"n_expert_used": 1,
|
| 11 |
+
"min_experts_per_node": 6
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [],
|
| 14 |
+
"mmproj": null,
|
| 15 |
+
"id": "glogwa68/Llama-4-scout-GGUF:Q4_K_M"
|
| 16 |
+
}
|
models/unsloth%2FDeepSeek-R1-Distill-Qwen-14B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "DeepSeek-R1-Distill-Qwen-14B-Q4_K_M",
|
| 3 |
+
"file": "DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf",
|
| 5 |
+
"size": "9.0GB",
|
| 6 |
+
"description": "DeepSeek R1 reasoning distilled into Qwen 14B",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FDeepSeek-R1-Distill-Qwen-32B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "DeepSeek-R1-Distill-Qwen-32B-Q4_K_M",
|
| 3 |
+
"file": "DeepSeek-R1-Distill-Qwen-32B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-32B-Q4_K_M.gguf",
|
| 5 |
+
"size": "19.9GB",
|
| 6 |
+
"description": "DeepSeek R1 reasoning distilled into Qwen 32B",
|
| 7 |
+
"draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FDevstral-Small-2505-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Devstral-Small-2505-Q4_K_M",
|
| 3 |
+
"file": "Devstral-Small-2505-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Devstral-Small-2505-GGUF/resolve/main/Devstral-Small-2505-Q4_K_M.gguf",
|
| 5 |
+
"size": "14.3GB",
|
| 6 |
+
"description": "Mistral agentic coding, tool use",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Devstral-Small-2505-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FGLM-4-32B-0414-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "GLM-4-32B-0414-Q4_K_M",
|
| 3 |
+
"file": "GLM-4-32B-0414-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/GLM-4-32B-0414-GGUF/resolve/main/GLM-4-32B-0414-Q4_K_M.gguf",
|
| 5 |
+
"size": "19.7GB",
|
| 6 |
+
"description": "Strong 32B, good tool calling",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/GLM-4-32B-0414-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FGLM-4.7-Flash-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "GLM-4.7-Flash-Q4_K_M",
|
| 3 |
+
"file": "GLM-4.7-Flash-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/GLM-4.7-Flash-GGUF/resolve/main/GLM-4.7-Flash-Q4_K_M.gguf",
|
| 5 |
+
"size": "18GB",
|
| 6 |
+
"description": "MoE 30B/3B active, 64 experts top-4, fast inference, tool calling",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 64,
|
| 10 |
+
"n_expert_used": 4,
|
| 11 |
+
"min_experts_per_node": 24
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [],
|
| 14 |
+
"mmproj": null,
|
| 15 |
+
"id": "unsloth/GLM-4.7-Flash-GGUF:Q4_K_M"
|
| 16 |
+
}
|
models/unsloth%2FMiniMax-M2.5-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "MiniMax-M2.5-Q4_K_M",
|
| 3 |
+
"file": "MiniMax-M2.5-Q4_K_M-00001-of-00004.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00001-of-00004.gguf",
|
| 5 |
+
"size": "138GB",
|
| 6 |
+
"description": "MiniMax-M2.5 MoE 456B/46B active, 256 experts top-8, Q4_K_M",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 256,
|
| 10 |
+
"n_expert_used": 8,
|
| 11 |
+
"min_experts_per_node": 96
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [
|
| 14 |
+
{
|
| 15 |
+
"file": "MiniMax-M2.5-Q4_K_M-00002-of-00004.gguf",
|
| 16 |
+
"url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00002-of-00004.gguf"
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"file": "MiniMax-M2.5-Q4_K_M-00003-of-00004.gguf",
|
| 20 |
+
"url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00003-of-00004.gguf"
|
| 21 |
+
},
|
| 22 |
+
{
|
| 23 |
+
"file": "MiniMax-M2.5-Q4_K_M-00004-of-00004.gguf",
|
| 24 |
+
"url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00004-of-00004.gguf"
|
| 25 |
+
}
|
| 26 |
+
],
|
| 27 |
+
"mmproj": null,
|
| 28 |
+
"id": "unsloth/MiniMax-M2.5-GGUF:Q4_K_M"
|
| 29 |
+
}
|
models/unsloth%2FMistral-Small-3.1-24B-Instruct-2503-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Mistral-Small-3.1-24B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF/resolve/main/Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf",
|
| 5 |
+
"size": "14.3GB",
|
| 6 |
+
"description": "Mistral general chat, good tool calling",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FQwen3-0.6B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-0.6B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-0.6B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-0.6B-GGUF/resolve/main/Qwen3-0.6B-Q4_K_M.gguf",
|
| 5 |
+
"size": "397MB",
|
| 6 |
+
"description": "Draft for Qwen3 models",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Qwen3-0.6B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FQwen3-14B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-14B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-14B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-14B-GGUF/resolve/main/Qwen3-14B-Q4_K_M.gguf",
|
| 5 |
+
"size": "9.0GB",
|
| 6 |
+
"description": "Qwen3 strong chat, thinking modes",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Qwen3-14B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FQwen3-30B-A3B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-30B-A3B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-30B-A3B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-30B-A3B-GGUF/resolve/main/Qwen3-30B-A3B-Q4_K_M.gguf",
|
| 5 |
+
"size": "17.3GB",
|
| 6 |
+
"description": "MoE general chat, 128 experts top-8, thinking/non-thinking",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 128,
|
| 10 |
+
"n_expert_used": 8,
|
| 11 |
+
"min_experts_per_node": 46
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [],
|
| 14 |
+
"mmproj": null,
|
| 15 |
+
"id": "unsloth/Qwen3-30B-A3B-GGUF:Q4_K_M"
|
| 16 |
+
}
|
models/unsloth%2FQwen3-32B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-32B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-32B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-32B-GGUF/resolve/main/Qwen3-32B-Q4_K_M.gguf",
|
| 5 |
+
"size": "19.8GB",
|
| 6 |
+
"description": "Best Qwen3 dense, thinking/non-thinking modes",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Qwen3-32B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FQwen3-4B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-4B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-4B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-4B-GGUF/resolve/main/Qwen3-4B-Q4_K_M.gguf",
|
| 5 |
+
"size": "2.5GB",
|
| 6 |
+
"description": "Qwen3 starter, thinking/non-thinking modes",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Qwen3-4B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FQwen3-8B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-8B-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-8B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-8B-GGUF/resolve/main/Qwen3-8B-Q4_K_M.gguf",
|
| 5 |
+
"size": "5.0GB",
|
| 6 |
+
"description": "Qwen3 mid-tier, strong for its size",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/Qwen3-8B-GGUF:Q4_K_M"
|
| 12 |
+
}
|
models/unsloth%2FQwen3-Coder-30B-A3B-Instruct-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M",
|
| 3 |
+
"file": "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF/resolve/main/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf",
|
| 5 |
+
"size": "18.6GB",
|
| 6 |
+
"description": "MoE agentic coding, tool use, 128 experts top-8",
|
| 7 |
+
"draft": "Qwen3-0.6B-Q4_K_M",
|
| 8 |
+
"moe": {
|
| 9 |
+
"n_expert": 128,
|
| 10 |
+
"n_expert_used": 8,
|
| 11 |
+
"min_experts_per_node": 46
|
| 12 |
+
},
|
| 13 |
+
"extra_files": [],
|
| 14 |
+
"mmproj": null,
|
| 15 |
+
"id": "unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF:Q4_K_M"
|
| 16 |
+
}
|
models/unsloth%2FQwen3.5-0.8B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3.5-0.8B-Vision-Q4_K_M",
|
| 3 |
+
"file": "Qwen3.5-0.8B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-0.8B-GGUF/resolve/main/Qwen3.5-0.8B-Q4_K_M.gguf",
|
| 5 |
+
"size": "508MB",
|
| 6 |
+
"description": "Tiny vision model, OCR, screenshots, runs anywhere",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": {
|
| 11 |
+
"file": "Qwen3.5-0.8B-mmproj-BF16.gguf",
|
| 12 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-0.8B-GGUF/resolve/main/mmproj-BF16.gguf"
|
| 13 |
+
},
|
| 14 |
+
"id": "unsloth/Qwen3.5-0.8B-GGUF:Q4_K_M"
|
| 15 |
+
}
|
models/unsloth%2FQwen3.5-4B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3.5-4B-Vision-Q4_K_M",
|
| 3 |
+
"file": "Qwen3.5-4B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-4B-GGUF/resolve/main/Qwen3.5-4B-Q4_K_M.gguf",
|
| 5 |
+
"size": "2.7GB",
|
| 6 |
+
"description": "Small vision model, good quality/size balance",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": {
|
| 11 |
+
"file": "Qwen3.5-4B-mmproj-BF16.gguf",
|
| 12 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-4B-GGUF/resolve/main/mmproj-BF16.gguf"
|
| 13 |
+
},
|
| 14 |
+
"id": "unsloth/Qwen3.5-4B-GGUF:Q4_K_M"
|
| 15 |
+
}
|
models/unsloth%2FQwen3.5-9B-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Qwen3.5-9B-Vision-Q4_K_M",
|
| 3 |
+
"file": "Qwen3.5-9B-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-9B-GGUF/resolve/main/Qwen3.5-9B-Q4_K_M.gguf",
|
| 5 |
+
"size": "5.8GB",
|
| 6 |
+
"description": "Vision + text, replaces Qwen3-8B with image understanding",
|
| 7 |
+
"draft": null,
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": {
|
| 11 |
+
"file": "Qwen3.5-9B-mmproj-BF16.gguf",
|
| 12 |
+
"url": "https://huggingface.co/unsloth/Qwen3.5-9B-GGUF/resolve/main/mmproj-BF16.gguf"
|
| 13 |
+
},
|
| 14 |
+
"id": "unsloth/Qwen3.5-9B-GGUF:Q4_K_M"
|
| 15 |
+
}
|
models/unsloth%2Fgemma-3-12b-it-GGUF%3AQ4_K_M/metadata.json
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"name": "Gemma-3-12B-it-Q4_K_M",
|
| 3 |
+
"file": "Gemma-3-12B-it-Q4_K_M.gguf",
|
| 4 |
+
"url": "https://huggingface.co/unsloth/gemma-3-12b-it-GGUF/resolve/main/gemma-3-12b-it-Q4_K_M.gguf",
|
| 5 |
+
"size": "7.3GB",
|
| 6 |
+
"description": "Google Gemma 3 12B, punches above weight",
|
| 7 |
+
"draft": "Gemma-3-1B-it-Q4_K_M",
|
| 8 |
+
"moe": null,
|
| 9 |
+
"extra_files": [],
|
| 10 |
+
"mmproj": null,
|
| 11 |
+
"id": "unsloth/gemma-3-12b-it-GGUF:Q4_K_M"
|
| 12 |
+
}
|