jamesdumay commited on
Commit
37dad18
·
verified ·
1 Parent(s): be6b08a

Migrate recommended model metadata to structured schema

Browse files

Rewrite recommended model metadata files to schema_version 1 with source, artifacts, and curation sections.

Files changed (39) hide show
  1. models/sha256-014f6b5897f35cf403936ded449f84395e5457a86756d3fc256c275cbb10c616/metadata.json +21 -9
  2. models/sha256-0302e8e8b4197e056d6bc9c962194ce93e23bfd5d4e23c0db201b7ca0f24acda/metadata.json +21 -9
  3. models/sha256-0afa2bbd49bcdec5d8de0e4e963e2ff05dcdeb77874aaa17037fc3a46298be97/metadata.json +26 -11
  4. models/sha256-124ed90efc924f62fc551abd896fa67010d7c901bf1db993c0b892f794ad4d1f/metadata.json +26 -11
  5. models/sha256-126ebe8597ded548b81baee6ca32cf1bec40b0bd8f16bfb242a3a445e6c1450c/metadata.json +21 -9
  6. models/sha256-141c18cb77bf3c8e6f0e1ed343fbb2eded986d0f9539f5931cd2cddcfbe0512b/metadata.json +21 -9
  7. models/sha256-142399797e62c91c1438748b84825409bf896bf1e006bd371857f03f1b5a1bb0/metadata.json +21 -9
  8. models/sha256-25c2acc8a85d230146810a47f1197d16adf712bf4da08ff9d652fed044a4f34d/metadata.json +21 -9
  9. models/sha256-2da8f5374c6ac42743cc97eb75c0f94ff38148f422bb5cc913871389b9cf70cd/metadata.json +21 -9
  10. models/sha256-30ce70b96bc1cbc708930855598c54daede525c5a6e548b1726e6f9f1ed40b60/metadata.json +21 -9
  11. models/sha256-317cc74e9dbc7b789d51088ce3430dafe566bd0dc6e72ab35d89f9cc90a5467e/metadata.json +35 -18
  12. models/sha256-3552e3d1b162204c88f4a5e194981156ac65141be87f9894b5b36835101979c1/metadata.json +21 -9
  13. models/sha256-38714a4ddb1c654b07df0e2bdde4acad0e5090cb54e656650f98927e2b4f3dbb/metadata.json +21 -9
  14. models/sha256-3ce9e41bcc7ac4942bd4237961306f9c6f0d2ad25ad1dca671acd65531450f7a/metadata.json +21 -9
  15. models/sha256-43da15b27fbddeb09b735ca73044097e55ea781bf9ced03cd45aedb09b20a428/metadata.json +21 -9
  16. models/sha256-530064106f7af37e49f2746a9ad5a84a59590621d441ad88d87a984a477c5cfe/metadata.json +21 -9
  17. models/sha256-60ae689aa5ccbd1387fd50d945ab13fd7936de5e0a90d5ba4358acd5a94fb44e/metadata.json +21 -9
  18. models/sha256-62449ec67d46ff80d8e87ae2f8e365af363d77ddea29c0ec1d612cdd98532992/metadata.json +21 -9
  19. models/sha256-65429236385af48afc1a32646b4c9f7f2ec596b6038b8c1df465251243454b31/metadata.json +26 -11
  20. models/sha256-654a3b00b6f00d880c15f4d6cd79d49e3c0940c6d9afb5b53feeb640a11b2612/metadata.json +21 -9
  21. models/sha256-6f94eb62fa4e5ea39ac7a9211fb5b0745039d373c715de4d074d59e3dd8601ac/metadata.json +21 -9
  22. models/sha256-74e53e5cd3017a27a921730edd97a09e41d7e32d5a7e151b18dbcd06dd63ee61/metadata.json +21 -9
  23. models/sha256-77f8a16681e07845a08e2adf70003d7743643daa8ec11f185f9b679b64252e4c/metadata.json +21 -9
  24. models/sha256-782855a0d5aef42bea734bc6a8d3e9c8b97c2d0b89ef1d77c337bc2c3b813b63/metadata.json +21 -9
  25. models/sha256-7c188c456ad2a0c9fcb17e661bf1e9c3197b8eb5ceaf08021ee060d9cd64d457/metadata.json +21 -9
  26. models/sha256-8b7b80a6b046a9b54d9c294dc01beb37ddb4e7fe06c66d4cca6d7c5b9659caf6/metadata.json +21 -9
  27. models/sha256-8e2a4b228397412bca5df002bf07f2aa24ba0f4f7c5afa4c99e1062635c975d6/metadata.json +21 -9
  28. models/sha256-9cd1d145bdc387fe992f48e08b4efebe148c943cb3a47ad346278ae5ee2d8681/metadata.json +21 -9
  29. models/sha256-9e44a1d1869f7bb5a5e639ba7afa1b6e4893e16d07176c3f60c59abc11e0bae5/metadata.json +21 -9
  30. models/sha256-a36bfb1bb182c377676973b1558376811757f5da9cd08ed0807ba968fb496b0c/metadata.json +21 -9
  31. models/sha256-adaf5e3d4b1ba78d51cf5d45f980e79ce57d068d1636e58552bbe596cf5acdef/metadata.json +21 -9
  32. models/sha256-bbedc72de1c38566d2f66e565a9e5a12038404da58b1eb952baf126f0d0a39b6/metadata.json +32 -15
  33. models/sha256-c3febba3f550e6c86ef95e079aefd50365718e213343feafe554a59b1cd3c42a/metadata.json +26 -11
  34. models/sha256-ce097731219bfb95091ea1767631e2c718b1585e3392353326fe24c8e7ec1bcb/metadata.json +21 -9
  35. models/sha256-d756c518cc3ec3c992152cd03514e5f98c0fb62880667f27716c877b3bebb44e/metadata.json +21 -9
  36. models/sha256-e49cac04ce4667accc61655e189035615c78711769aa7b8a43fdca887b366436/metadata.json +21 -9
  37. models/sha256-e72a7ff194d35e87832281741d2d4bba639d8d5913c55a3aa9e945a6fcfb7c37/metadata.json +21 -9
  38. models/sha256-f488ea7c68e65fdd7775f04d58766601ee1a7d8f571a154c25f2467bb295f369/metadata.json +21 -9
  39. models/sha256-feb131c258a0c95b5f36109f87b2d0f04462916ce5e2197f129d888c7e0b2a4a/metadata.json +21 -9
models/sha256-014f6b5897f35cf403936ded449f84395e5457a86756d3fc256c275cbb10c616/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen3-4B-Q4_K_M",
3
- "file": "Qwen3-4B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-4B-GGUF/resolve/main/Qwen3-4B-Q4_K_M.gguf",
5
- "size": "2.5GB",
6
- "description": "Qwen3 starter, thinking/non-thinking modes",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Qwen3-4B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-4B-GGUF:Q4_K_M",
4
  "name": "Qwen3-4B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-4B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-4B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-4B-GGUF/resolve/main/Qwen3-4B-Q4_K_M.gguf",
16
+ "size_bytes": 2500000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Qwen3 starter, thinking/non-thinking modes",
21
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-0302e8e8b4197e056d6bc9c962194ce93e23bfd5d4e23c0db201b7ca0f24acda/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Llama-3.2-1B-Instruct-Q4_K_M",
3
- "file": "Llama-3.2-1B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF/resolve/main/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
5
- "size": "760MB",
6
- "description": "Draft for Llama 3.x and Llama 4 models",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/Llama-3.2-1B-Instruct-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/Llama-3.2-1B-Instruct-GGUF:Q4_K_M",
4
  "name": "Llama-3.2-1B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/Llama-3.2-1B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Llama-3.2-1B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF/resolve/main/Llama-3.2-1B-Instruct-Q4_K_M.gguf",
16
+ "size_bytes": 760000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Draft for Llama 3.x and Llama 4 models",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-0afa2bbd49bcdec5d8de0e4e963e2ff05dcdeb77874aaa17037fc3a46298be97/metadata.json CHANGED
@@ -1,15 +1,30 @@
1
  {
 
 
2
  "name": "Qwen3.5-27B-Q4_K_M",
3
- "file": "Qwen3.5-27B-Q4_K_M.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/qwen3.5/blobs/sha256:d4b8b4f4c350f5d322dc8235175eeae02d32c6f3fd70bdb9ea481e3abb7d7fc4",
5
- "size": "17GB",
6
- "description": "Qwen3.5 27B, vision + text, strong reasoning and coding",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": {
11
- "file": "Qwen3.5-27B-mmproj-BF16.gguf",
12
- "url": "https://huggingface.co/unsloth/Qwen3.5-27B-GGUF/resolve/main/mmproj-BF16.gguf"
13
  },
14
- "id": "Qwen3.5-27B-Q4_K_M"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen3.5-27B-Q4_K_M",
4
  "name": "Qwen3.5-27B-Q4_K_M",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
 
 
 
 
 
10
  },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3.5-27B-Q4_K_M.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/qwen3.5/blobs/sha256:d4b8b4f4c350f5d322dc8235175eeae02d32c6f3fd70bdb9ea481e3abb7d7fc4",
16
+ "size_bytes": 17000000000
17
+ },
18
+ {
19
+ "role": "mmproj",
20
+ "path": "mmproj-BF16.gguf",
21
+ "url": "https://huggingface.co/unsloth/Qwen3.5-27B-GGUF/resolve/main/mmproj-BF16.gguf",
22
+ "size_bytes": null
23
+ }
24
+ ],
25
+ "curation": {
26
+ "description": "Qwen3.5 27B, vision + text, strong reasoning and coding",
27
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
28
+ },
29
+ "moe": null
30
  }
models/sha256-124ed90efc924f62fc551abd896fa67010d7c901bf1db993c0b892f794ad4d1f/metadata.json CHANGED
@@ -1,15 +1,30 @@
1
  {
 
 
2
  "name": "Qwen3.5-0.8B-Vision-Q4_K_M",
3
- "file": "Qwen3.5-0.8B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3.5-0.8B-GGUF/resolve/main/Qwen3.5-0.8B-Q4_K_M.gguf",
5
- "size": "508MB",
6
- "description": "Tiny vision model, OCR, screenshots, runs anywhere",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": {
11
- "file": "Qwen3.5-0.8B-mmproj-BF16.gguf",
12
- "url": "https://huggingface.co/unsloth/Qwen3.5-0.8B-GGUF/resolve/main/mmproj-BF16.gguf"
13
  },
14
- "id": "unsloth/Qwen3.5-0.8B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3.5-0.8B-GGUF:Q4_K_M",
4
  "name": "Qwen3.5-0.8B-Vision-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3.5-0.8B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
 
 
 
 
 
10
  },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3.5-0.8B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3.5-0.8B-GGUF/resolve/main/Qwen3.5-0.8B-Q4_K_M.gguf",
16
+ "size_bytes": 508000000
17
+ },
18
+ {
19
+ "role": "mmproj",
20
+ "path": "mmproj-BF16.gguf",
21
+ "url": "https://huggingface.co/unsloth/Qwen3.5-0.8B-GGUF/resolve/main/mmproj-BF16.gguf",
22
+ "size_bytes": null
23
+ }
24
+ ],
25
+ "curation": {
26
+ "description": "Tiny vision model, OCR, screenshots, runs anywhere",
27
+ "draft_model_id": null
28
+ },
29
+ "moe": null
30
  }
models/sha256-126ebe8597ded548b81baee6ca32cf1bec40b0bd8f16bfb242a3a445e6c1450c/metadata.json CHANGED
@@ -1,16 +1,28 @@
1
  {
 
 
2
  "name": "GLM-4.7-Flash-Q4_K_M",
3
- "file": "GLM-4.7-Flash-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/GLM-4.7-Flash-GGUF/resolve/main/GLM-4.7-Flash-Q4_K_M.gguf",
5
- "size": "18GB",
6
- "description": "MoE 30B/3B active, 64 experts top-4, fast inference, tool calling",
7
- "draft": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "moe": {
9
  "n_expert": 64,
10
  "n_expert_used": 4,
11
  "min_experts_per_node": 24
12
- },
13
- "extra_files": [],
14
- "mmproj": null,
15
- "id": "unsloth/GLM-4.7-Flash-GGUF:Q4_K_M"
16
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/GLM-4.7-Flash-GGUF:Q4_K_M",
4
  "name": "GLM-4.7-Flash-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/GLM-4.7-Flash-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "GLM-4.7-Flash-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/GLM-4.7-Flash-GGUF/resolve/main/GLM-4.7-Flash-Q4_K_M.gguf",
16
+ "size_bytes": 18000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "MoE 30B/3B active, 64 experts top-4, fast inference, tool calling",
21
+ "draft_model_id": null
22
+ },
23
  "moe": {
24
  "n_expert": 64,
25
  "n_expert_used": 4,
26
  "min_experts_per_node": 24
27
+ }
 
 
 
28
  }
models/sha256-141c18cb77bf3c8e6f0e1ed343fbb2eded986d0f9539f5931cd2cddcfbe0512b/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "DeepSeek-R1-Distill-Qwen-14B-Q4_K_M",
3
- "file": "DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf",
5
- "size": "9.0GB",
6
- "description": "DeepSeek R1 reasoning distilled into Qwen 14B",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF:Q4_K_M",
4
  "name": "DeepSeek-R1-Distill-Qwen-14B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-14B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf",
16
+ "size_bytes": 9000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "DeepSeek R1 reasoning distilled into Qwen 14B",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-142399797e62c91c1438748b84825409bf896bf1e006bd371857f03f1b5a1bb0/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-32B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-32B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/Qwen2.5-32B-Instruct-GGUF/resolve/main/Qwen2.5-32B-Instruct-Q4_K_M.gguf",
5
- "size": "20GB",
6
- "description": "Proven general chat",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/Qwen2.5-32B-Instruct-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/Qwen2.5-32B-Instruct-GGUF:Q4_K_M",
4
  "name": "Qwen2.5-32B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/Qwen2.5-32B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen2.5-32B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/Qwen2.5-32B-Instruct-GGUF/resolve/main/Qwen2.5-32B-Instruct-Q4_K_M.gguf",
16
+ "size_bytes": 20000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Proven general chat",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-25c2acc8a85d230146810a47f1197d16adf712bf4da08ff9d652fed044a4f34d/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-Coder-7B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-Coder-7B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct-GGUF/resolve/main/qwen2.5-coder-7b-instruct-q4_k_m.gguf",
5
- "size": "4.4GB",
6
- "description": "Code generation & completion",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Qwen/Qwen2.5-Coder-7B-Instruct-GGUF/qwen2.5-coder-7b-instruct-q4_k_m.gguf"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen/Qwen2.5-Coder-7B-Instruct-GGUF/qwen2.5-coder-7b-instruct-q4_k_m.gguf",
4
  "name": "Qwen2.5-Coder-7B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "Qwen/Qwen2.5-Coder-7B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": null
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "qwen2.5-coder-7b-instruct-q4_k_m.gguf",
15
+ "url": "https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct-GGUF/resolve/main/qwen2.5-coder-7b-instruct-q4_k_m.gguf",
16
+ "size_bytes": 4400000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Code generation & completion",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-2da8f5374c6ac42743cc97eb75c0f94ff38148f422bb5cc913871389b9cf70cd/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-72B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-72B-Instruct-Q4_K_M.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/qwen2.5/blobs/sha256:6e7fdda508e91cb0f63de5c15ff79ac63a1584ccafd751c07ca12b7f442101b8",
5
- "size": "47GB",
6
- "description": "Flagship Qwen2.5, great tensor split showcase",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Qwen2.5-72B-Instruct-Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen2.5-72B-Instruct-Q4_K_M",
4
  "name": "Qwen2.5-72B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen2.5-72B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/qwen2.5/blobs/sha256:6e7fdda508e91cb0f63de5c15ff79ac63a1584ccafd751c07ca12b7f442101b8",
16
+ "size_bytes": 47000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Flagship Qwen2.5, great tensor split showcase",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-30ce70b96bc1cbc708930855598c54daede525c5a6e548b1726e6f9f1ed40b60/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen3-14B-Q4_K_M",
3
- "file": "Qwen3-14B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-14B-GGUF/resolve/main/Qwen3-14B-Q4_K_M.gguf",
5
- "size": "9.0GB",
6
- "description": "Qwen3 strong chat, thinking modes",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Qwen3-14B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-14B-GGUF:Q4_K_M",
4
  "name": "Qwen3-14B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-14B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-14B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-14B-GGUF/resolve/main/Qwen3-14B-Q4_K_M.gguf",
16
+ "size_bytes": 9000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Qwen3 strong chat, thinking modes",
21
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-317cc74e9dbc7b789d51088ce3430dafe566bd0dc6e72ab35d89f9cc90a5467e/metadata.json CHANGED
@@ -1,29 +1,46 @@
1
  {
 
 
2
  "name": "MiniMax-M2.5-Q4_K_M",
3
- "file": "MiniMax-M2.5-Q4_K_M-00001-of-00004.gguf",
4
- "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00001-of-00004.gguf",
5
- "size": "138GB",
6
- "description": "MiniMax-M2.5 MoE 456B/46B active, 256 experts top-8, Q4_K_M",
7
- "draft": null,
8
- "moe": {
9
- "n_expert": 256,
10
- "n_expert_used": 8,
11
- "min_experts_per_node": 96
12
  },
13
- "extra_files": [
 
 
 
 
 
 
14
  {
15
- "file": "MiniMax-M2.5-Q4_K_M-00002-of-00004.gguf",
16
- "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00002-of-00004.gguf"
 
 
17
  },
18
  {
19
- "file": "MiniMax-M2.5-Q4_K_M-00003-of-00004.gguf",
20
- "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00003-of-00004.gguf"
 
 
21
  },
22
  {
23
- "file": "MiniMax-M2.5-Q4_K_M-00004-of-00004.gguf",
24
- "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00004-of-00004.gguf"
 
 
25
  }
26
  ],
27
- "mmproj": null,
28
- "id": "unsloth/MiniMax-M2.5-GGUF:Q4_K_M"
 
 
 
 
 
 
 
29
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/MiniMax-M2.5-GGUF:Q4_K_M",
4
  "name": "MiniMax-M2.5-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/MiniMax-M2.5-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
 
 
 
 
10
  },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Q4_K_M/MiniMax-M2.5-Q4_K_M-00001-of-00004.gguf",
15
+ "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00001-of-00004.gguf",
16
+ "size_bytes": 138000000000
17
+ },
18
  {
19
+ "role": "split",
20
+ "path": "Q4_K_M/MiniMax-M2.5-Q4_K_M-00002-of-00004.gguf",
21
+ "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00002-of-00004.gguf",
22
+ "size_bytes": null
23
  },
24
  {
25
+ "role": "split",
26
+ "path": "Q4_K_M/MiniMax-M2.5-Q4_K_M-00003-of-00004.gguf",
27
+ "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00003-of-00004.gguf",
28
+ "size_bytes": null
29
  },
30
  {
31
+ "role": "split",
32
+ "path": "Q4_K_M/MiniMax-M2.5-Q4_K_M-00004-of-00004.gguf",
33
+ "url": "https://huggingface.co/unsloth/MiniMax-M2.5-GGUF/resolve/main/Q4_K_M/MiniMax-M2.5-Q4_K_M-00004-of-00004.gguf",
34
+ "size_bytes": null
35
  }
36
  ],
37
+ "curation": {
38
+ "description": "MiniMax-M2.5 MoE 456B/46B active, 256 experts top-8, Q4_K_M",
39
+ "draft_model_id": null
40
+ },
41
+ "moe": {
42
+ "n_expert": 256,
43
+ "n_expert_used": 8,
44
+ "min_experts_per_node": 96
45
+ }
46
  }
models/sha256-3552e3d1b162204c88f4a5e194981156ac65141be87f9894b5b36835101979c1/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-Coder-14B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-Coder-14B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/Qwen2.5-Coder-14B-Instruct-GGUF/resolve/main/Qwen2.5-Coder-14B-Instruct-Q4_K_M.gguf",
5
- "size": "9.0GB",
6
- "description": "Strong code gen, fills gap between 7B and 32B",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/Qwen2.5-Coder-14B-Instruct-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/Qwen2.5-Coder-14B-Instruct-GGUF:Q4_K_M",
4
  "name": "Qwen2.5-Coder-14B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/Qwen2.5-Coder-14B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen2.5-Coder-14B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/Qwen2.5-Coder-14B-Instruct-GGUF/resolve/main/Qwen2.5-Coder-14B-Instruct-Q4_K_M.gguf",
16
+ "size_bytes": 9000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Strong code gen, fills gap between 7B and 32B",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-38714a4ddb1c654b07df0e2bdde4acad0e5090cb54e656650f98927e2b4f3dbb/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Gemma-3-12B-it-Q4_K_M",
3
- "file": "Gemma-3-12B-it-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/gemma-3-12b-it-GGUF/resolve/main/gemma-3-12b-it-Q4_K_M.gguf",
5
- "size": "7.3GB",
6
- "description": "Google Gemma 3 12B, punches above weight",
7
- "draft": "Gemma-3-1B-it-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/gemma-3-12b-it-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/gemma-3-12b-it-GGUF:Q4_K_M",
4
  "name": "Gemma-3-12B-it-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/gemma-3-12b-it-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "gemma-3-12b-it-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/gemma-3-12b-it-GGUF/resolve/main/gemma-3-12b-it-Q4_K_M.gguf",
16
+ "size_bytes": 7300000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Google Gemma 3 12B, punches above weight",
21
+ "draft_model_id": "Gemma-3-1B-it-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-3ce9e41bcc7ac4942bd4237961306f9c6f0d2ad25ad1dca671acd65531450f7a/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Mistral-Small-3.1-24B-Instruct-Q4_K_M",
3
- "file": "Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF/resolve/main/Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf",
5
- "size": "14.3GB",
6
- "description": "Mistral general chat, good tool calling",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF:Q4_K_M",
4
  "name": "Mistral-Small-3.1-24B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Mistral-Small-3.1-24B-Instruct-2503-GGUF/resolve/main/Mistral-Small-3.1-24B-Instruct-2503-Q4_K_M.gguf",
16
+ "size_bytes": 14300000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Mistral general chat, good tool calling",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-43da15b27fbddeb09b735ca73044097e55ea781bf9ced03cd45aedb09b20a428/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Gemma-3-1B-it-Q4_K_M",
3
- "file": "Gemma-3-1B-it-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/google_gemma-3-1b-it-GGUF/resolve/main/google_gemma-3-1b-it-Q4_K_M.gguf",
5
- "size": "780MB",
6
- "description": "Draft for Gemma 3 models",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/google_gemma-3-1b-it-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/google_gemma-3-1b-it-GGUF:Q4_K_M",
4
  "name": "Gemma-3-1B-it-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/google_gemma-3-1b-it-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "google_gemma-3-1b-it-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/google_gemma-3-1b-it-GGUF/resolve/main/google_gemma-3-1b-it-Q4_K_M.gguf",
16
+ "size_bytes": 780000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Draft for Gemma 3 models",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-530064106f7af37e49f2746a9ad5a84a59590621d441ad88d87a984a477c5cfe/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-14B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-14B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/Qwen2.5-14B-Instruct-GGUF/resolve/main/Qwen2.5-14B-Instruct-Q4_K_M.gguf",
5
- "size": "9.0GB",
6
- "description": "Solid general chat",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/Qwen2.5-14B-Instruct-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/Qwen2.5-14B-Instruct-GGUF:Q4_K_M",
4
  "name": "Qwen2.5-14B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/Qwen2.5-14B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen2.5-14B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/Qwen2.5-14B-Instruct-GGUF/resolve/main/Qwen2.5-14B-Instruct-Q4_K_M.gguf",
16
+ "size_bytes": 9000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Solid general chat",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-60ae689aa5ccbd1387fd50d945ab13fd7936de5e0a90d5ba4358acd5a94fb44e/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-0.5B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-0.5B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/qwen2.5-0.5b-instruct-q4_k_m.gguf",
5
- "size": "491MB",
6
- "description": "Draft for Qwen2.5 and DeepSeek-R1-Distill models",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF/qwen2.5-0.5b-instruct-q4_k_m.gguf"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen/Qwen2.5-0.5B-Instruct-GGUF/qwen2.5-0.5b-instruct-q4_k_m.gguf",
4
  "name": "Qwen2.5-0.5B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "Qwen/Qwen2.5-0.5B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": null
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "qwen2.5-0.5b-instruct-q4_k_m.gguf",
15
+ "url": "https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct-GGUF/resolve/main/qwen2.5-0.5b-instruct-q4_k_m.gguf",
16
+ "size_bytes": 491000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Draft for Qwen2.5 and DeepSeek-R1-Distill models",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-62449ec67d46ff80d8e87ae2f8e365af363d77ddea29c0ec1d612cdd98532992/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Hermes-2-Pro-Mistral-7B-Q4_K_M",
3
- "file": "Hermes-2-Pro-Mistral-7B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/Hermes-2-Pro-Mistral-7B-GGUF/resolve/main/Hermes-2-Pro-Mistral-7B-Q4_K_M.gguf",
5
- "size": "4.4GB",
6
- "description": "Goose default, strong tool calling for agents",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/Hermes-2-Pro-Mistral-7B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/Hermes-2-Pro-Mistral-7B-GGUF:Q4_K_M",
4
  "name": "Hermes-2-Pro-Mistral-7B-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/Hermes-2-Pro-Mistral-7B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Hermes-2-Pro-Mistral-7B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/Hermes-2-Pro-Mistral-7B-GGUF/resolve/main/Hermes-2-Pro-Mistral-7B-Q4_K_M.gguf",
16
+ "size_bytes": 4400000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Goose default, strong tool calling for agents",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-65429236385af48afc1a32646b4c9f7f2ec596b6038b8c1df465251243454b31/metadata.json CHANGED
@@ -1,15 +1,30 @@
1
  {
 
 
2
  "name": "Qwen3.5-9B-Vision-Q4_K_M",
3
- "file": "Qwen3.5-9B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3.5-9B-GGUF/resolve/main/Qwen3.5-9B-Q4_K_M.gguf",
5
- "size": "5.8GB",
6
- "description": "Vision + text, replaces Qwen3-8B with image understanding",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": {
11
- "file": "Qwen3.5-9B-mmproj-BF16.gguf",
12
- "url": "https://huggingface.co/unsloth/Qwen3.5-9B-GGUF/resolve/main/mmproj-BF16.gguf"
13
  },
14
- "id": "unsloth/Qwen3.5-9B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3.5-9B-GGUF:Q4_K_M",
4
  "name": "Qwen3.5-9B-Vision-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3.5-9B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
 
 
 
 
 
10
  },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3.5-9B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3.5-9B-GGUF/resolve/main/Qwen3.5-9B-Q4_K_M.gguf",
16
+ "size_bytes": 5800000000
17
+ },
18
+ {
19
+ "role": "mmproj",
20
+ "path": "mmproj-BF16.gguf",
21
+ "url": "https://huggingface.co/unsloth/Qwen3.5-9B-GGUF/resolve/main/mmproj-BF16.gguf",
22
+ "size_bytes": null
23
+ }
24
+ ],
25
+ "curation": {
26
+ "description": "Vision + text, replaces Qwen3-8B with image understanding",
27
+ "draft_model_id": null
28
+ },
29
+ "moe": null
30
  }
models/sha256-654a3b00b6f00d880c15f4d6cd79d49e3c0940c6d9afb5b53feeb640a11b2612/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Llama-3.3-70B-Instruct-Q4_K_M",
3
- "file": "Llama-3.3-70B-Instruct-Q4_K_M.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/llama3.3/blobs/sha256:4824460d29f2058aaf6e1118a63a7a197a09bed509f0e7d4e2efb1ee273b447d",
5
- "size": "43GB",
6
- "description": "Meta Llama 3.3 70B, strong all-around",
7
- "draft": "Llama-3.2-1B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Llama-3.3-70B-Instruct-Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Llama-3.3-70B-Instruct-Q4_K_M",
4
  "name": "Llama-3.3-70B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Llama-3.3-70B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/llama3.3/blobs/sha256:4824460d29f2058aaf6e1118a63a7a197a09bed509f0e7d4e2efb1ee273b447d",
16
+ "size_bytes": 43000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Meta Llama 3.3 70B, strong all-around",
21
+ "draft_model_id": "Llama-3.2-1B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-6f94eb62fa4e5ea39ac7a9211fb5b0745039d373c715de4d074d59e3dd8601ac/metadata.json CHANGED
@@ -1,16 +1,28 @@
1
  {
 
 
2
  "name": "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M",
3
- "file": "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF/resolve/main/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf",
5
- "size": "18.6GB",
6
- "description": "MoE agentic coding, tool use, 128 experts top-8",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "moe": {
9
  "n_expert": 128,
10
  "n_expert_used": 8,
11
  "min_experts_per_node": 46
12
- },
13
- "extra_files": [],
14
- "mmproj": null,
15
- "id": "unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF:Q4_K_M"
16
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF:Q4_K_M",
4
  "name": "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF/resolve/main/Qwen3-Coder-30B-A3B-Instruct-Q4_K_M.gguf",
16
+ "size_bytes": 18600000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "MoE agentic coding, tool use, 128 experts top-8",
21
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
22
+ },
23
  "moe": {
24
  "n_expert": 128,
25
  "n_expert_used": 8,
26
  "min_experts_per_node": 46
27
+ }
 
 
 
28
  }
models/sha256-74e53e5cd3017a27a921730edd97a09e41d7e32d5a7e151b18dbcd06dd63ee61/metadata.json CHANGED
@@ -1,16 +1,28 @@
1
  {
 
 
2
  "name": "Mixtral-8x22B-Instruct-Q4_K_M",
3
- "file": "Mixtral-8x22B-Instruct-Q4_K_M.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/mixtral/blobs/sha256:f3329ad0c787f4f73cab99e8c877bb76403060561dd0caa318127683c87bbcb4",
5
- "size": "86GB",
6
- "description": "Mixtral 8x22B MoE, 8 experts top-2",
7
- "draft": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "moe": {
9
  "n_expert": 8,
10
  "n_expert_used": 2,
11
  "min_experts_per_node": 4
12
- },
13
- "extra_files": [],
14
- "mmproj": null,
15
- "id": "Mixtral-8x22B-Instruct-Q4_K_M"
16
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Mixtral-8x22B-Instruct-Q4_K_M",
4
  "name": "Mixtral-8x22B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Mixtral-8x22B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/mixtral/blobs/sha256:f3329ad0c787f4f73cab99e8c877bb76403060561dd0caa318127683c87bbcb4",
16
+ "size_bytes": 86000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Mixtral 8x22B MoE, 8 experts top-2",
21
+ "draft_model_id": null
22
+ },
23
  "moe": {
24
  "n_expert": 8,
25
  "n_expert_used": 2,
26
  "min_experts_per_node": 4
27
+ }
 
 
 
28
  }
models/sha256-77f8a16681e07845a08e2adf70003d7743643daa8ec11f185f9b679b64252e4c/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen3-32B-Q4_K_M",
3
- "file": "Qwen3-32B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-32B-GGUF/resolve/main/Qwen3-32B-Q4_K_M.gguf",
5
- "size": "19.8GB",
6
- "description": "Best Qwen3 dense, thinking/non-thinking modes",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Qwen3-32B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-32B-GGUF:Q4_K_M",
4
  "name": "Qwen3-32B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-32B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-32B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-32B-GGUF/resolve/main/Qwen3-32B-Q4_K_M.gguf",
16
+ "size_bytes": 19800000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Best Qwen3 dense, thinking/non-thinking modes",
21
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-782855a0d5aef42bea734bc6a8d3e9c8b97c2d0b89ef1d77c337bc2c3b813b63/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen3-0.6B-Q4_K_M",
3
- "file": "Qwen3-0.6B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-0.6B-GGUF/resolve/main/Qwen3-0.6B-Q4_K_M.gguf",
5
- "size": "397MB",
6
- "description": "Draft for Qwen3 models",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Qwen3-0.6B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-0.6B-GGUF:Q4_K_M",
4
  "name": "Qwen3-0.6B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-0.6B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-0.6B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-0.6B-GGUF/resolve/main/Qwen3-0.6B-Q4_K_M.gguf",
16
+ "size_bytes": 397000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Draft for Qwen3 models",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-7c188c456ad2a0c9fcb17e661bf1e9c3197b8eb5ceaf08021ee060d9cd64d457/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "DeepSeek-R1-Distill-70B-Q4_K_M",
3
- "file": "DeepSeek-R1-Distill-70B-Q4_K_M.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/deepseek-r1/blobs/sha256:4cd576d9aa16961244012223abf01445567b061f1814b57dfef699e4cf8df339",
5
- "size": "43GB",
6
- "description": "DeepSeek R1 distilled to 70B (Qwen2.5-based), strong reasoning",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "DeepSeek-R1-Distill-70B-Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "DeepSeek-R1-Distill-70B-Q4_K_M",
4
  "name": "DeepSeek-R1-Distill-70B-Q4_K_M",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "DeepSeek-R1-Distill-70B-Q4_K_M.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/deepseek-r1/blobs/sha256:4cd576d9aa16961244012223abf01445567b061f1814b57dfef699e4cf8df339",
16
+ "size_bytes": 43000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "DeepSeek R1 distilled to 70B (Qwen2.5-based), strong reasoning",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-8b7b80a6b046a9b54d9c294dc01beb37ddb4e7fe06c66d4cca6d7c5b9659caf6/metadata.json CHANGED
@@ -1,16 +1,28 @@
1
  {
 
 
2
  "name": "Qwen3-235B-A22B-Q4_K_M",
3
- "file": "Qwen3-235B-A22B-Q4_K_M.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/qwen3/blobs/sha256:aeacdadecbed8a07e42026d1a1d3cd30715bb2994ebe4e4ca4009e1a4abe8d5d",
5
- "size": "142GB",
6
- "description": "Qwen3 235B MoE A22B active, 128 experts top-8",
7
- "draft": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "moe": {
9
  "n_expert": 128,
10
  "n_expert_used": 8,
11
  "min_experts_per_node": 46
12
- },
13
- "extra_files": [],
14
- "mmproj": null,
15
- "id": "Qwen3-235B-A22B-Q4_K_M"
16
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen3-235B-A22B-Q4_K_M",
4
  "name": "Qwen3-235B-A22B-Q4_K_M",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-235B-A22B-Q4_K_M.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/qwen3/blobs/sha256:aeacdadecbed8a07e42026d1a1d3cd30715bb2994ebe4e4ca4009e1a4abe8d5d",
16
+ "size_bytes": 142000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Qwen3 235B MoE A22B active, 128 experts top-8",
21
+ "draft_model_id": null
22
+ },
23
  "moe": {
24
  "n_expert": 128,
25
  "n_expert_used": 8,
26
  "min_experts_per_node": 46
27
+ }
 
 
 
28
  }
models/sha256-8e2a4b228397412bca5df002bf07f2aa24ba0f4f7c5afa4c99e1062635c975d6/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "DeepSeek-R1-Distill-Qwen-32B-Q4_K_M",
3
- "file": "DeepSeek-R1-Distill-Qwen-32B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-32B-Q4_K_M.gguf",
5
- "size": "19.9GB",
6
- "description": "DeepSeek R1 reasoning distilled into Qwen 32B",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF:Q4_K_M",
4
  "name": "DeepSeek-R1-Distill-Qwen-32B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "DeepSeek-R1-Distill-Qwen-32B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-32B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-32B-Q4_K_M.gguf",
16
+ "size_bytes": 19900000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "DeepSeek R1 reasoning distilled into Qwen 32B",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-9cd1d145bdc387fe992f48e08b4efebe148c943cb3a47ad346278ae5ee2d8681/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Devstral-Small-2505-Q4_K_M",
3
- "file": "Devstral-Small-2505-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Devstral-Small-2505-GGUF/resolve/main/Devstral-Small-2505-Q4_K_M.gguf",
5
- "size": "14.3GB",
6
- "description": "Mistral agentic coding, tool use",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Devstral-Small-2505-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Devstral-Small-2505-GGUF:Q4_K_M",
4
  "name": "Devstral-Small-2505-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Devstral-Small-2505-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Devstral-Small-2505-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Devstral-Small-2505-GGUF/resolve/main/Devstral-Small-2505-Q4_K_M.gguf",
16
+ "size_bytes": 14300000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Mistral agentic coding, tool use",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-9e44a1d1869f7bb5a5e639ba7afa1b6e4893e16d07176c3f60c59abc11e0bae5/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Gemma-3-27B-it-Q4_K_M",
3
- "file": "Gemma-3-27B-it-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/google_gemma-3-27b-it-GGUF/resolve/main/google_gemma-3-27b-it-Q4_K_M.gguf",
5
- "size": "17GB",
6
- "description": "Google Gemma 3 27B, strong reasoning",
7
- "draft": "Gemma-3-1B-it-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/google_gemma-3-27b-it-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/google_gemma-3-27b-it-GGUF:Q4_K_M",
4
  "name": "Gemma-3-27B-it-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/google_gemma-3-27b-it-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "google_gemma-3-27b-it-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/google_gemma-3-27b-it-GGUF/resolve/main/google_gemma-3-27b-it-Q4_K_M.gguf",
16
+ "size_bytes": 17000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Google Gemma 3 27B, strong reasoning",
21
+ "draft_model_id": "Gemma-3-1B-it-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-a36bfb1bb182c377676973b1558376811757f5da9cd08ed0807ba968fb496b0c/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-3B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-3B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct-GGUF/resolve/main/qwen2.5-3b-instruct-q4_k_m.gguf",
5
- "size": "2.1GB",
6
- "description": "Small & fast general chat",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Qwen/Qwen2.5-3B-Instruct-GGUF/qwen2.5-3b-instruct-q4_k_m.gguf"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen/Qwen2.5-3B-Instruct-GGUF/qwen2.5-3b-instruct-q4_k_m.gguf",
4
  "name": "Qwen2.5-3B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "Qwen/Qwen2.5-3B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": null
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "qwen2.5-3b-instruct-q4_k_m.gguf",
15
+ "url": "https://huggingface.co/Qwen/Qwen2.5-3B-Instruct-GGUF/resolve/main/qwen2.5-3b-instruct-q4_k_m.gguf",
16
+ "size_bytes": 2100000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Small & fast general chat",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-adaf5e3d4b1ba78d51cf5d45f980e79ce57d068d1636e58552bbe596cf5acdef/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Llama-3.2-3B-Instruct-Q4_K_M",
3
- "file": "Llama-3.2-3B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_M.gguf",
5
- "size": "2.0GB",
6
- "description": "Meta Llama 3.2, goose default, good tool calling",
7
- "draft": "Llama-3.2-1B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M",
4
  "name": "Llama-3.2-3B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "bartowski/Llama-3.2-3B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Llama-3.2-3B-Instruct-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_M.gguf",
16
+ "size_bytes": 2000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Meta Llama 3.2, goose default, good tool calling",
21
+ "draft_model_id": "Llama-3.2-1B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-bbedc72de1c38566d2f66e565a9e5a12038404da58b1eb952baf126f0d0a39b6/metadata.json CHANGED
@@ -1,25 +1,42 @@
1
  {
 
 
2
  "name": "Qwen3-Coder-Next-Q4_K_M",
3
- "file": "Qwen3-Coder-Next-Q4_K_M-00001-of-00004.gguf",
4
- "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00001-of-00004.gguf",
5
- "size": "48GB",
6
- "description": "Qwen3 Coder Next ~85B dense, frontier coding model",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [
10
  {
11
- "file": "Qwen3-Coder-Next-Q4_K_M-00002-of-00004.gguf",
12
- "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00002-of-00004.gguf"
 
 
13
  },
14
  {
15
- "file": "Qwen3-Coder-Next-Q4_K_M-00003-of-00004.gguf",
16
- "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00003-of-00004.gguf"
 
 
17
  },
18
  {
19
- "file": "Qwen3-Coder-Next-Q4_K_M-00004-of-00004.gguf",
20
- "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00004-of-00004.gguf"
 
 
 
 
 
 
 
 
21
  }
22
  ],
23
- "mmproj": null,
24
- "id": "Qwen/Qwen3-Coder-Next-GGUF:Qwen3-Coder-Next-Q4_K_M"
 
 
 
25
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen/Qwen3-Coder-Next-GGUF:Qwen3-Coder-Next-Q4_K_M",
4
  "name": "Qwen3-Coder-Next-Q4_K_M",
5
+ "source": {
6
+ "repo": "Qwen/Qwen3-Coder-Next-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Qwen3-Coder-Next-Q4_K_M"
10
+ },
11
+ "artifacts": [
12
  {
13
+ "role": "primary",
14
+ "path": "Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00001-of-00004.gguf",
15
+ "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00001-of-00004.gguf",
16
+ "size_bytes": 48000000000
17
  },
18
  {
19
+ "role": "split",
20
+ "path": "Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00002-of-00004.gguf",
21
+ "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00002-of-00004.gguf",
22
+ "size_bytes": null
23
  },
24
  {
25
+ "role": "split",
26
+ "path": "Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00003-of-00004.gguf",
27
+ "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00003-of-00004.gguf",
28
+ "size_bytes": null
29
+ },
30
+ {
31
+ "role": "split",
32
+ "path": "Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00004-of-00004.gguf",
33
+ "url": "https://huggingface.co/Qwen/Qwen3-Coder-Next-GGUF/resolve/main/Qwen3-Coder-Next-Q4_K_M/Qwen3-Coder-Next-Q4_K_M-00004-of-00004.gguf",
34
+ "size_bytes": null
35
  }
36
  ],
37
+ "curation": {
38
+ "description": "Qwen3 Coder Next ~85B dense, frontier coding model",
39
+ "draft_model_id": null
40
+ },
41
+ "moe": null
42
  }
models/sha256-c3febba3f550e6c86ef95e079aefd50365718e213343feafe554a59b1cd3c42a/metadata.json CHANGED
@@ -1,15 +1,30 @@
1
  {
 
 
2
  "name": "Qwen3.5-4B-Vision-Q4_K_M",
3
- "file": "Qwen3.5-4B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3.5-4B-GGUF/resolve/main/Qwen3.5-4B-Q4_K_M.gguf",
5
- "size": "2.7GB",
6
- "description": "Small vision model, good quality/size balance",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": {
11
- "file": "Qwen3.5-4B-mmproj-BF16.gguf",
12
- "url": "https://huggingface.co/unsloth/Qwen3.5-4B-GGUF/resolve/main/mmproj-BF16.gguf"
13
  },
14
- "id": "unsloth/Qwen3.5-4B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3.5-4B-GGUF:Q4_K_M",
4
  "name": "Qwen3.5-4B-Vision-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3.5-4B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
 
 
 
 
 
10
  },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3.5-4B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3.5-4B-GGUF/resolve/main/Qwen3.5-4B-Q4_K_M.gguf",
16
+ "size_bytes": 2700000000
17
+ },
18
+ {
19
+ "role": "mmproj",
20
+ "path": "mmproj-BF16.gguf",
21
+ "url": "https://huggingface.co/unsloth/Qwen3.5-4B-GGUF/resolve/main/mmproj-BF16.gguf",
22
+ "size_bytes": null
23
+ }
24
+ ],
25
+ "curation": {
26
+ "description": "Small vision model, good quality/size balance",
27
+ "draft_model_id": null
28
+ },
29
+ "moe": null
30
  }
models/sha256-ce097731219bfb95091ea1767631e2c718b1585e3392353326fe24c8e7ec1bcb/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen3-8B-Q4_K_M",
3
- "file": "Qwen3-8B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-8B-GGUF/resolve/main/Qwen3-8B-Q4_K_M.gguf",
5
- "size": "5.0GB",
6
- "description": "Qwen3 mid-tier, strong for its size",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/Qwen3-8B-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-8B-GGUF:Q4_K_M",
4
  "name": "Qwen3-8B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-8B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-8B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-8B-GGUF/resolve/main/Qwen3-8B-Q4_K_M.gguf",
16
+ "size_bytes": 5000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Qwen3 mid-tier, strong for its size",
21
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-d756c518cc3ec3c992152cd03514e5f98c0fb62880667f27716c877b3bebb44e/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Qwen2.5-Coder-32B-Instruct-Q4_K_M",
3
- "file": "Qwen2.5-Coder-32B-Instruct-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct-GGUF/resolve/main/qwen2.5-coder-32b-instruct-q4_k_m.gguf",
5
- "size": "20GB",
6
- "description": "Top-tier code gen, matches GPT-4o on code",
7
- "draft": "Qwen2.5-0.5B-Instruct-Q4_K_M",
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Qwen/Qwen2.5-Coder-32B-Instruct-GGUF/qwen2.5-coder-32b-instruct-q4_k_m.gguf"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Qwen/Qwen2.5-Coder-32B-Instruct-GGUF/qwen2.5-coder-32b-instruct-q4_k_m.gguf",
4
  "name": "Qwen2.5-Coder-32B-Instruct-Q4_K_M",
5
+ "source": {
6
+ "repo": "Qwen/Qwen2.5-Coder-32B-Instruct-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": null
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "qwen2.5-coder-32b-instruct-q4_k_m.gguf",
15
+ "url": "https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct-GGUF/resolve/main/qwen2.5-coder-32b-instruct-q4_k_m.gguf",
16
+ "size_bytes": 20000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Top-tier code gen, matches GPT-4o on code",
21
+ "draft_model_id": "Qwen2.5-0.5B-Instruct-Q4_K_M"
22
+ },
23
+ "moe": null
24
  }
models/sha256-e49cac04ce4667accc61655e189035615c78711769aa7b8a43fdca887b366436/metadata.json CHANGED
@@ -1,16 +1,28 @@
1
  {
 
 
2
  "name": "Llama-4-Scout-Q4_K_M",
3
- "file": "Llama-4-Scout-4bit-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/glogwa68/Llama-4-scout-GGUF/resolve/main/Llama-4-Scout-4bit-Q4_K_M.gguf",
5
- "size": "22.5GB",
6
- "description": "MoE 109B/17B active, 16 experts top-1, Meta latest, tool calling",
7
- "draft": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "moe": {
9
  "n_expert": 16,
10
  "n_expert_used": 1,
11
  "min_experts_per_node": 6
12
- },
13
- "extra_files": [],
14
- "mmproj": null,
15
- "id": "glogwa68/Llama-4-scout-GGUF:Q4_K_M"
16
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "glogwa68/Llama-4-scout-GGUF:Q4_K_M",
4
  "name": "Llama-4-Scout-Q4_K_M",
5
+ "source": {
6
+ "repo": "glogwa68/Llama-4-scout-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Llama-4-Scout-4bit-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/glogwa68/Llama-4-scout-GGUF/resolve/main/Llama-4-Scout-4bit-Q4_K_M.gguf",
16
+ "size_bytes": 22500000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "MoE 109B/17B active, 16 experts top-1, Meta latest, tool calling",
21
+ "draft_model_id": null
22
+ },
23
  "moe": {
24
  "n_expert": 16,
25
  "n_expert_used": 1,
26
  "min_experts_per_node": 6
27
+ }
 
 
 
28
  }
models/sha256-e72a7ff194d35e87832281741d2d4bba639d8d5913c55a3aa9e945a6fcfb7c37/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "Llama-3.1-405B-Instruct-Q2_K",
3
- "file": "Llama-3.1-405B-Instruct-Q2_K.gguf",
4
- "url": "https://registry.ollama.ai/v2/library/llama3.1/blobs/sha256:e7e1972e5b13caead8a8dd9c94f4a0dec59ac2d9dd52e0cd1c067e6077eb4677",
5
- "size": "149GB",
6
- "description": "Llama 3.1 405B Instruct Q2_K, largest dense model",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "Llama-3.1-405B-Instruct-Q2_K"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "Llama-3.1-405B-Instruct-Q2_K",
4
  "name": "Llama-3.1-405B-Instruct-Q2_K",
5
+ "source": {
6
+ "repo": "",
7
+ "revision": null,
8
+ "commit": null,
9
+ "selector": "Q2_K"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Llama-3.1-405B-Instruct-Q2_K.gguf",
15
+ "url": "https://registry.ollama.ai/v2/library/llama3.1/blobs/sha256:e7e1972e5b13caead8a8dd9c94f4a0dec59ac2d9dd52e0cd1c067e6077eb4677",
16
+ "size_bytes": 149000000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Llama 3.1 405B Instruct Q2_K, largest dense model",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }
models/sha256-f488ea7c68e65fdd7775f04d58766601ee1a7d8f571a154c25f2467bb295f369/metadata.json CHANGED
@@ -1,16 +1,28 @@
1
  {
 
 
2
  "name": "Qwen3-30B-A3B-Q4_K_M",
3
- "file": "Qwen3-30B-A3B-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/Qwen3-30B-A3B-GGUF/resolve/main/Qwen3-30B-A3B-Q4_K_M.gguf",
5
- "size": "17.3GB",
6
- "description": "MoE general chat, 128 experts top-8, thinking/non-thinking",
7
- "draft": "Qwen3-0.6B-Q4_K_M",
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "moe": {
9
  "n_expert": 128,
10
  "n_expert_used": 8,
11
  "min_experts_per_node": 46
12
- },
13
- "extra_files": [],
14
- "mmproj": null,
15
- "id": "unsloth/Qwen3-30B-A3B-GGUF:Q4_K_M"
16
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/Qwen3-30B-A3B-GGUF:Q4_K_M",
4
  "name": "Qwen3-30B-A3B-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/Qwen3-30B-A3B-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "Qwen3-30B-A3B-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/Qwen3-30B-A3B-GGUF/resolve/main/Qwen3-30B-A3B-Q4_K_M.gguf",
16
+ "size_bytes": 17300000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "MoE general chat, 128 experts top-8, thinking/non-thinking",
21
+ "draft_model_id": "Qwen3-0.6B-Q4_K_M"
22
+ },
23
  "moe": {
24
  "n_expert": 128,
25
  "n_expert_used": 8,
26
  "min_experts_per_node": 46
27
+ }
 
 
 
28
  }
models/sha256-feb131c258a0c95b5f36109f87b2d0f04462916ce5e2197f129d888c7e0b2a4a/metadata.json CHANGED
@@ -1,12 +1,24 @@
1
  {
 
 
2
  "name": "GLM-4-32B-0414-Q4_K_M",
3
- "file": "GLM-4-32B-0414-Q4_K_M.gguf",
4
- "url": "https://huggingface.co/unsloth/GLM-4-32B-0414-GGUF/resolve/main/GLM-4-32B-0414-Q4_K_M.gguf",
5
- "size": "19.7GB",
6
- "description": "Strong 32B, good tool calling",
7
- "draft": null,
8
- "moe": null,
9
- "extra_files": [],
10
- "mmproj": null,
11
- "id": "unsloth/GLM-4-32B-0414-GGUF:Q4_K_M"
 
 
 
 
 
 
 
 
 
 
12
  }
 
1
  {
2
+ "schema_version": 1,
3
+ "id": "unsloth/GLM-4-32B-0414-GGUF:Q4_K_M",
4
  "name": "GLM-4-32B-0414-Q4_K_M",
5
+ "source": {
6
+ "repo": "unsloth/GLM-4-32B-0414-GGUF",
7
+ "revision": "main",
8
+ "commit": null,
9
+ "selector": "Q4_K_M"
10
+ },
11
+ "artifacts": [
12
+ {
13
+ "role": "primary",
14
+ "path": "GLM-4-32B-0414-Q4_K_M.gguf",
15
+ "url": "https://huggingface.co/unsloth/GLM-4-32B-0414-GGUF/resolve/main/GLM-4-32B-0414-Q4_K_M.gguf",
16
+ "size_bytes": 19700000000
17
+ }
18
+ ],
19
+ "curation": {
20
+ "description": "Strong 32B, good tool calling",
21
+ "draft_model_id": null
22
+ },
23
+ "moe": null
24
  }