File size: 2,065 Bytes
602967f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
---
license: apache-2.0
language:
- en
library_name: transformers
tags:
- code
- python
- maincoder
- code-generation
- gguf
- quantized
pipeline_tag: text-generation
base_model: Maincode/Maincoder-1B
---
<img src="https://huggingface.co/datasets/Maincode/assets/resolve/e51154e034201be1a5dad0e9c8de31d8b9f17643/maincoder_logo.png" alt="" width="1250">

# Maincoder-1B-GGUF

GGUF quantizations of [**Maincoder-1B**](https://huggingface.co/Maincode/Maincoder-1B), a code-focused language model optimized for code generation and completion tasks. These quantized versions are designed for efficient local deployment with [llama.cpp](https://github.com/ggerganov/llama.cpp).

Find more details in the original model card: https://huggingface.co/Maincode/Maincoder-1B

## How to run Maincoder

Example usage with llama.cpp:

```bash
llama-cli -hf Maincode/Maincoder-1B-GGUF
```

Or with a specific quantization:

```bash
llama-cli -hf Maincode/Maincoder-1B-GGUF -m Maincoder-1B-Q4_K_M.gguf
```

Code completion example:

```bash
llama-cli -hf Maincode/Maincoder-1B-GGUF -p 'def fibonacci(n: int) -> int:
    """Return the n-th Fibonacci number."""
' -n 256
```

## Available Quantizations

| Filename | Size | Description |
|----------|------|-------------|
| Maincoder-1B-BF16.gguf | 1.9 GB | BFloat16 - Full precision, best quality |
| Maincoder-1B-F16.gguf | 1.9 GB | Float16 - Full precision |
| Maincoder-1B-Q8_0.gguf | 1.0 GB | 8-bit quantization - Highest quality quantized |
| Maincoder-1B-Q6_K.gguf | 809 MB | 6-bit quantization - High quality |
| Maincoder-1B-Q5_K_M.gguf | 722 MB | 5-bit quantization - Great balance |
| Maincoder-1B-Q4_K_M.gguf | 641 MB | 4-bit quantization - Recommended |
| Maincoder-1B-Q4_0.gguf | 614 MB | 4-bit quantization - Smallest, fastest |

## 📄 License

This model is released under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0).

## 🔗 Links

- [Original Model](https://huggingface.co/Maincode/Maincoder-1B)
- [Maincode](https://maincode.com)
- [llama.cpp](https://github.com/ggerganov/llama.cpp)