File size: 3,869 Bytes
947084c 6afff48 947084c b64c479 2bfced4 947084c b64c479 947084c 913e55c 2bfced4 efe1705 2bfced4 913e55c 2bfced4 c2491c3 2bfced4 751dc21 913e55c 2bfced4 efe1705 2bfced4 0de06ca 2bfced4 913e55c 2bfced4 913e55c 2bfced4 913e55c 947084c 0de06ca 947084c 913e55c 947084c 913e55c bdca4ff 947084c 6afff48 947084c 8f39e71 947084c 8f39e71 e76abfb 947084c b28a07b c9dc71b 522c183 4a82d52 c9dc71b 522c183 10a76f9 c9dc71b 10a76f9 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 | ---
tags:
- gguf-connector
widget:
- text: a cat in a hat
output:
url: https://raw.githubusercontent.com/calcuis/gguf-pack/master/w8g.png
- text: a raccoon in a hat
output:
url: https://raw.githubusercontent.com/calcuis/gguf-pack/master/w8f.png
- text: a raccoon in a hat
output:
url: https://raw.githubusercontent.com/calcuis/gguf-pack/master/w6a.png
- text: a dog walking in a cyber city with joy
output:
url: https://raw.githubusercontent.com/calcuis/gguf-pack/master/w6b.png
- text: a dog walking in a cyber city with joy
output:
url: https://raw.githubusercontent.com/calcuis/gguf-pack/master/w6c.png
- text: a dog walking in a cyber city with joy
output:
url: https://raw.githubusercontent.com/calcuis/gguf-pack/master/w8e.png
---
## self-hosted api
- run it with `gguf-connector`; activate the backend in console/terminal by
```
ggc w8
```
- choose your model* file
>
>GGUF available. Select which one to use:
>
>1. sd3.5-2b-lite-iq4_nl.gguf [[1.74GB](https://huggingface.co/calcuis/sd3.5-lite-gguf/blob/main/sd3.5-2b-lite-iq4_nl.gguf)]
>2. sd3.5-2b-lite-mxfp4_moe.gguf [[2.86GB](https://huggingface.co/calcuis/sd3.5-lite-gguf/blob/main/sd3.5-2b-lite-mxfp4_moe.gguf)]
>
>Enter your choice (1 to 2): _
>
*accept sd3.5 2b model gguf recently, this will give you the fastest experience for even low tier gpu; frontend https://test.gguf.org or localhost (see **decentralized frontend** section below)

- or opt fastapi **lumina** connector
```
ggc w7
```
- choose your model* file
>
>GGUF available. Select which one to use:
>
>1. lumina2-q4_0.gguf [[1.47GB](https://huggingface.co/calcuis/lumina-gguf/blob/main/lumina2-q4_0.gguf)]
>2. lumina2-q8_0.gguf [[2.77GB](https://huggingface.co/calcuis/lumina-gguf/blob/main/lumina2-q8_0.gguf)]
>
>Enter your choice (1 to 2): _
>
*as lumina is no lite version recently, might need to increase the step to around 25 for better output
- or opt fastapi **flux** connector
```
ggc w6
```
- choose your model* file
>
>GGUF available. Select which one to use:
>
>1. flux-dev-lite-q2_k.gguf [[4.08GB](https://huggingface.co/calcuis/krea-gguf/blob/main/flux-dev-lite-q2_k.gguf)]
>2. flux-krea-lite-q2_k.gguf [[4.08GB](https://huggingface.co/calcuis/krea-gguf/blob/main/flux-krea-lite-q2_k.gguf)]
>
>Enter your choice (1 to 2): _
>
*accept any flux model gguf, lite is recommended for saving loading time

- flexible frontend choice (see below)
## decentralized frontend
- option 1: navigate to https://test.gguf.org

- option 2: localhost; keep the backend running and open a new terminal session then execute
```
ggc b
```

<Gallery />
## self-hosted api (edit)
- run it with `gguf-connector`; activate the backend in console/terminal by
```
ggc e8
```
- choose your model file
>
>GGUF available. Select which one to use:
>
>1. flux-kontext-lite-q2_k.gguf [[4.08GB](https://huggingface.co/calcuis/kontext-gguf/blob/main/flux-kontext-lite-q2_k.gguf)]
>
>Enter your choice (1 to 1): _
>

## decentralized frontend - opt `Edit` from pulldown menu (exclusive for 🐷 holder recently)
- option 1: navigate to https://gguf.org


- option 2: localhost; keep the backend running and open a new terminal session then execute
```
ggc a
```
 |