uaytug commited on
Commit
d837104
·
verified ·
1 Parent(s): 49bec63

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md CHANGED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # uCoder-8b-base-GGUF
2
+
3
+ Quantized GGUF models converted from [uaytug/uCoder-8b-base](https://huggingface.co/uaytug/uCoder-8b-base).
4
+
5
+ Converted using the latest llama.cpp (CUDA-accelerated quantization).
6
+
7
+ ### Available Files
8
+
9
+ **16-bit**
10
+ - `uCoder-8b-base-BF16.gguf` → Highest precision float (similar to original, ~16 GB)
11
+
12
+ **8-bit**
13
+ - `uCoder-8b-base-Q8_0.gguf` → Near-lossless
14
+
15
+ **6-bit**
16
+ - `uCoder-8b-base-Q6_K.gguf`
17
+
18
+ **5-bit**
19
+ - `uCoder-8b-base-Q5_K_S.gguf`
20
+ - `uCoder-8b-base-Q5_K_M.gguf` → Great quality
21
+
22
+ **4-bit** (most popular range)
23
+ - `uCoder-8b-base-Q4_K_M.gguf` → **Recommended balance**
24
+ - `uCoder-8b-base-Q4_K_S.gguf`
25
+ - `uCoder-8b-base-Q4_1.gguf`
26
+
27
+ **3-bit**
28
+ - `uCoder-8b-base-Q3_K_S.gguf`
29
+ - `uCoder-8b-base-Q3_K_M.gguf`
30
+
31
+ **2-bit**
32
+ - `uCoder-8b-base-Q2_K.gguf`
33
+
34
+ ### Usage
35
+ Use with llama.cpp, LM Studio, Ollama, etc.
36
+
37
+ This is a **base** model — best for code/further fine-tuning.