Xin Liu commited on
Commit
b79de6a
·
1 Parent(s): 84e5a5e

Signed-off-by: Xin Liu <sam@secondstate.io>

Files changed (1) hide show
  1. README.md +76 -15
README.md CHANGED
@@ -1,15 +1,76 @@
1
- # Deepseek-Coder-6.7B-Instruct
2
-
3
- ## Prompt Template
4
-
5
- ```text
6
- You are an AI programming assistant, utilizing the DeepSeek Coder model, developed by DeepSeek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.
7
- ### Instruction:
8
- {question_1}
9
- ### Response:
10
- {answer_1}
11
- <|EOT|>
12
- ### Instruction:
13
- {question_2}
14
- ### Response:
15
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: deepseek-ai/deepseek-coder-6.7b-instruct
3
+ inference: false
4
+ license: other
5
+ license_link: LICENSE
6
+ license_name: deepseek
7
+ model_creator: DeepSeek
8
+ model_name: Deepseek Coder 6.7B Instruct
9
+ model_type: deepseek
10
+ quantized_by: Second State Inc.
11
+ ---
12
+
13
+ <!-- header start -->
14
+ <!-- 200823 -->
15
+ <div style="width: auto; margin-left: auto; margin-right: auto">
16
+ <img src="https://github.com/second-state/LlamaEdge/raw/dev/assets/logo.svg" style="width: 100%; min-width: 400px; display: block; margin: auto;">
17
+ </div>
18
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
19
+ <!-- header end -->
20
+
21
+ # Deepseek-Coder-6.7B-Instruct-GGUF
22
+
23
+ ## Original Model
24
+
25
+ [deepseek-ai/deepseek-coder-6.7b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct)
26
+
27
+ ## Run with LlamaEdge
28
+
29
+ - LlamaEdge version: [v0.2.4](https://github.com/second-state/LlamaEdge/releases/tag/0.2.4)
30
+
31
+ - Prompt template
32
+
33
+ - Prompt type: `deepseek-coder`
34
+
35
+ - Prompt string
36
+
37
+ ```text
38
+ {system}
39
+ ### Instruction:
40
+ {question_1}
41
+ ### Response:
42
+ {answer_1}
43
+ <|EOT|>
44
+ ### Instruction:
45
+ {question_2}
46
+ ### Response:
47
+ ```
48
+
49
+ - Run as LlamaEdge service
50
+
51
+ ```bash
52
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:deepseek-coder-6.7b-instruct.Q5_K_M.gguf llama-api-server.wasm -p deepseek-coder
53
+ ```
54
+
55
+ - Run as LlamaEdge command app
56
+
57
+ ```bash
58
+ wasmedge --dir .:. --nn-preload default:GGML:AUTO:deepseek-coder-6.7b-instruct.Q5_K_M.gguf llama-chat.wasm -p deepseek-coder
59
+ ```
60
+
61
+ ## Quantized GGUF Models
62
+
63
+ | Name | Quant method | Bits | Size | Use case |
64
+ | ---- | ---- | ---- | ---- | ----- |
65
+ | [deepseek-coder-6.7b-instruct-Q2_K.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q2_K.gguf) | Q2_K | 2 | 2.53 GB| smallest, significant quality loss - not recommended for most purposes |
66
+ | [deepseek-coder-6.7b-instruct-Q3_K_L.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q3_K_L.gguf) | Q3_K_L | 3 | 3.6 GB| small, substantial quality loss |
67
+ | [deepseek-coder-6.7b-instruct-Q3_K_M.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q3_K_M.gguf) | Q3_K_M | 3 | 3.3 GB| very small, high quality loss |
68
+ | [deepseek-coder-6.7b-instruct-Q3_K_S.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q3_K_S.gguf) | Q3_K_S | 3 | 2.95 GB| very small, high quality loss |
69
+ | [deepseek-coder-6.7b-instruct-Q4_0.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q4_0.gguf) | Q4_0 | 4 | 3.83 GB| legacy; small, very high quality loss - prefer using Q3_K_M |
70
+ | [deepseek-coder-6.7b-instruct-Q4_K_M.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q4_K_M.gguf) | Q4_K_M | 4 | 4.08 GB| medium, balanced quality - recommended |
71
+ | [deepseek-coder-6.7b-instruct-Q4_K_S.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q4_K_S.gguf) | Q4_K_S | 4 | 3.86 GB| small, greater quality loss |
72
+ | [deepseek-coder-6.7b-instruct-Q5_0.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q5_0.gguf) | Q5_0 | 5 | 4.65 GB| legacy; medium, balanced quality - prefer using Q4_K_M |
73
+ | [deepseek-coder-6.7b-instruct-Q5_K_M.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q5_K_M.gguf) | Q5_K_M | 5 | 4.79 GB| large, very low quality loss - recommended |
74
+ | [deepseek-coder-6.7b-instruct-Q5_K_S.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q5_K_S.gguf) | Q5_K_S | 5 | 4.65 GB| large, low quality loss - recommended |
75
+ | [deepseek-coder-6.7b-instruct-Q6_K.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q6_K.gguf) | Q6_K | 6 | 5.53 GB| very large, extremely low quality loss |
76
+ | [deepseek-coder-6.7b-instruct-Q8_0.gguf](https://huggingface.co/second-state/Deepseek-Coder-6.7B-Instruct-GGUF/blob/main/deepseek-coder-6.7b-instruct-Q8_0.gguf) | Q8_0 | 8 | 7.16 GB| very large, extremely low quality loss - not recommended |