Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -102,11 +102,56 @@ quantized_by: Second State Inc.
|
|
| 102 |
| [DeepSeek-Coder-V2-Instruct-Q4_K_S-00003-of-00005.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q4_K_S-00003-of-00005.gguf) | Q4_K_S | 4 | 29.9 GB| small, greater quality loss |
|
| 103 |
| [DeepSeek-Coder-V2-Instruct-Q4_K_S-00004-of-00005.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q4_K_S-00004-of-00005.gguf) | Q4_K_S | 4 | 29.8 GB| small, greater quality loss |
|
| 104 |
| [DeepSeek-Coder-V2-Instruct-Q4_K_S-00005-of-00005.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q4_K_S-00005-of-00005.gguf) | Q4_K_S | 4 | 14.8 GB| small, greater quality loss |
|
| 105 |
-
| [DeepSeek-Coder-V2-Instruct-Q5_0.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0.gguf) | Q5_0 | 5 |
|
| 106 |
-
| [DeepSeek-Coder-V2-Instruct-
|
| 107 |
-
| [DeepSeek-Coder-V2-Instruct-
|
| 108 |
-
| [DeepSeek-Coder-V2-Instruct-
|
| 109 |
-
| [DeepSeek-Coder-V2-Instruct-
|
| 110 |
-
| [DeepSeek-Coder-V2-Instruct-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 111 |
|
| 112 |
*Quatized with llama.cpp b3499*
|
|
|
|
| 102 |
| [DeepSeek-Coder-V2-Instruct-Q4_K_S-00003-of-00005.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q4_K_S-00003-of-00005.gguf) | Q4_K_S | 4 | 29.9 GB| small, greater quality loss |
|
| 103 |
| [DeepSeek-Coder-V2-Instruct-Q4_K_S-00004-of-00005.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q4_K_S-00004-of-00005.gguf) | Q4_K_S | 4 | 29.8 GB| small, greater quality loss |
|
| 104 |
| [DeepSeek-Coder-V2-Instruct-Q4_K_S-00005-of-00005.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q4_K_S-00005-of-00005.gguf) | Q4_K_S | 4 | 14.8 GB| small, greater quality loss |
|
| 105 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_0-00001-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0-00001-of-00006.gguf) | Q5_0 | 5 | 29.4 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 106 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_0-00002-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0-00002-of-00006.gguf) | Q5_0 | 5 | 29.2 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 107 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_0-00003-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0-00003-of-00006.gguf) | Q5_0 | 5 | 30.0 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 108 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_0-00004-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0-00004-of-00006.gguf) | Q5_0 | 5 | 29.2 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 109 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_0-00005-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0-00005-of-00006.gguf) | Q5_0 | 5 | 29.2 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 110 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_0-00006-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_0-00006-of-00006.gguf) | Q5_0 | 5 | 15.4 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
|
| 111 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_M-00001-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_M-00001-of-00006.gguf) | Q5_K_M | 5 | 29.7 GB| large, very low quality loss - recommended |
|
| 112 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_M-00002-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_M-00002-of-00006.gguf) | Q5_K_M | 5 | 29.7 GB| large, very low quality loss - recommended |
|
| 113 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_M-00003-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_M-00003-of-00006.gguf) | Q5_K_M | 5 | 29.7 GB| large, very low quality loss - recommended |
|
| 114 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_M-00004-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_M-00004-of-00006.gguf) | Q5_K_M | 5 | 29.9 GB| large, very low quality loss - recommended |
|
| 115 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_M-00005-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_M-00005-of-00006.gguf) | Q5_K_M | 5 | 29.9 GB| large, very low quality loss - recommended |
|
| 116 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_M-00006-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_M-00006-of-00006.gguf) | Q5_K_M | 5 | 18.3 GB| large, very low quality loss - recommended |
|
| 117 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_S-00001-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_S-00001-of-00006.gguf) | Q5_K_S | 5 | 29.4 GB| large, low quality loss - recommended |
|
| 118 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_S-00002-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_S-00002-of-00006.gguf) | Q5_K_S | 5 | 29.2 GB| large, low quality loss - recommended |
|
| 119 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_S-00003-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_S-00003-of-00006.gguf) | Q5_K_S | 5 | 30.0 GB| large, low quality loss - recommended |
|
| 120 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_S-00004-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_S-00004-of-00006.gguf) | Q5_K_S | 5 | 29.2 GB| large, low quality loss - recommended |
|
| 121 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_S-00005-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_S-00005-of-00006.gguf) | Q5_K_S | 5 | 29.2 GB| large, low quality loss - recommended |
|
| 122 |
+
| [DeepSeek-Coder-V2-Instruct-Q5_K_S-00006-of-00006.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q5_K_S-00006-of-00006.gguf) | Q5_K_S | 5 | 15.4 GB| large, low quality loss - recommended |
|
| 123 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00001-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00001-of-00007.gguf) | Q6_K | 6 | 29.6 GB| very large, extremely low quality loss |
|
| 124 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00002-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00002-of-00007.gguf) | Q6_K | 6 | 29.0 GB| very large, extremely low quality loss |
|
| 125 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00003-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00003-of-00007.gguf) | Q6_K | 6 | 29.5 GB| very large, extremely low quality loss |
|
| 126 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00004-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00004-of-00007.gguf) | Q6_K | 6 | 29.3 GB| very large, extremely low quality loss |
|
| 127 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00005-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00005-of-00007.gguf) | Q6_K | 6 | 29.3 GB| very large, extremely low quality loss |
|
| 128 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00006-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00006-of-00007.gguf) | Q6_K | 6 | 29.3 GB| very large, extremely low quality loss |
|
| 129 |
+
| [DeepSeek-Coder-V2-Instruct-Q6_K-00007-of-00007.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q6_K-00007-of-00007.gguf) | Q6_K | 6 | 17.3 GB| very large, extremely low quality loss |
|
| 130 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00001-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00001-of-00009.gguf) | Q8_0 | 8 | 29.7 GB| very large, extremely low quality loss - not recommended |
|
| 131 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00002-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00002-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 132 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00003-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00003-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 133 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00004-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00004-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 134 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00005-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00005-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 135 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00006-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00006-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 136 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00007-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00007-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 137 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00008-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00008-of-00009.gguf) | Q8_0 | 8 | 29.6 GB| very large, extremely low quality loss - not recommended |
|
| 138 |
+
| [DeepSeek-Coder-V2-Instruct-Q8_0-00009-of-00009.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-Q8_0-00009-of-00009.gguf) | Q8_0 | 8 | 14.0 GB| very large, extremely low quality loss - not recommended |
|
| 139 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00001-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00001-of-00017.gguf) | f16 | 16 | 29.5 GB| |
|
| 140 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00002-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00002-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 141 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00003-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00003-of-00017.gguf) | f16 | 16 | 28.9 GB| |
|
| 142 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00004-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00004-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 143 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00005-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00005-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 144 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00006-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00006-of-00017.gguf) | f16 | 16 | 28.9 GB| |
|
| 145 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00007-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00007-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 146 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00008-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00008-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 147 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00009-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00009-of-00017.gguf) | f16 | 16 | 28.9 GB| |
|
| 148 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00010-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00010-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 149 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00011-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00011-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 150 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00012-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00012-of-00017.gguf) | f16 | 16 | 28.5 GB| |
|
| 151 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00013-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00013-of-00017.gguf) | f16 | 16 | 29.7 GB| |
|
| 152 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00014-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00014-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 153 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00015-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00015-of-00017.gguf) | f16 | 16 | 28.9 GB| |
|
| 154 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00016-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00016-of-00017.gguf) | f16 | 16 | 29.3 GB| |
|
| 155 |
+
| [DeepSeek-Coder-V2-Instruct-f16-00017-of-00017.gguf](https://huggingface.co/second-state/DeepSeek-Coder-V2-Instruct-GGUF/blob/main/DeepSeek-Coder-V2-Instruct-f16-00017-of-00017.gguf) | f16 | 16 | 5.03 GB| |
|
| 156 |
|
| 157 |
*Quatized with llama.cpp b3499*
|