File size: 2,108 Bytes
2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 080dc59 2d850e9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 |
---
license: apache-2.0
tags:
- gguf
- llama.cpp
- qwen
- uncensored
- quantized
- offline
- local-ai
---
# Qwen3 1.7B β Q8 GGUF (Uncensored, 32K Context)
This repository contains a **fully uncensored** and **quantized (Q8_0)** GGUF version of **Qwen3 1.7B**, designed for **offline, local inference** using `llama.cpp` and compatible runtimes.
By default, the model operates in **thinking mode**.
If you prefer a **non-thinking (direct) response mode**, simply add **`/no_think`** before your prompt.
- β
**Uncensored**
- β
**32K context length**
- β
**Q8_0 quantization**
- β
**Offline / local use**
- β
**No LoRA required (merged / base inference)**
---
## π Model Details
- **Base Model**: Qwen3 1.7B
- **Format**: GGUF
- **Quantization**: Q8_0
- **Context Length**: 32,000 tokens
- **Intended Use**:
- Offline assistants
- Email writing
- Small coding tasks
- Automation
- General daily usage
- **Not intended for**:
- Hosted public services
- Safety-restricted environments
---
## βΆοΈ Usage (llama.cpp)
```bash
./llama-cli \
-m gguf/qwen3-1.7b-q8_0.gguf \
-p "Hello"
```
# Recommended flags
```bash
--temp 0.2
--top-p 0.9
```
For concise outputs:
```text
Answer directly. Use yes or no when possible.
```
## β οΈ Disclaimer
- This model is **fully uncensored** and provided **as-is**.
- You are responsible for how you use it
- Do not deploy in public-facing applications without moderation
- Intended for **personal, research, and offline use**
## π§ Quantization Info
- **Q8_0** provides near-FP16 quality
- Stable outputs
- Recommended for CPU and mobile-class devices
## π€ Author & Organization
- **Creator**: Thirumalai
- **Company**: ZFusionAI
## π License
- Apache 2.0
---
## π― Final note
This README is:
- β
Honest (uncensored clearly stated)
- β
Clean for Hugging Face
- β
Professional (company + creator credited)
- β
No policy-bait wording
If you want, next I can:
- tighten it for **discoverability**
- add **benchmarks**
- or generate a **model card version**
You shipped this like a pro ππ₯
|