GTKING commited on
Commit
080dc59
Β·
verified Β·
1 Parent(s): d6f50e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +85 -39
README.md CHANGED
@@ -1,58 +1,104 @@
1
  ---
2
- base_model: meta-llama/Llama-3.2-1B-Instruct
3
- library_name: transformers
4
- model_name: ZFusionAI_Hacker
5
  tags:
6
- - generated_from_trainer
7
- - trl
8
- - sft
9
- licence: license
 
 
 
10
  ---
11
 
12
- # Model Card for ZFusionAI_Hacker
13
 
14
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct).
15
- It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
- ## Quick start
 
18
 
19
- ```python
20
- from transformers import pipeline
 
 
 
21
 
22
- question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
- generator = pipeline("text-generation", model="GTKING/ZFusionAI_Hacker", device="cuda")
24
- output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
- print(output["generated_text"])
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  ```
27
 
28
- ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/gamingking9025-muthayammal-college-of-engineering/huggingface/runs/ajpft44z)
 
 
 
31
 
 
32
 
33
- This model was trained with SFT.
 
 
34
 
35
- ### Framework versions
36
 
37
- - TRL: 0.26.2
38
- - Transformers: 4.57.1
39
- - Pytorch: 2.8.0+cu126
40
- - Datasets: 4.4.1
41
- - Tokenizers: 0.22.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
 
43
- ## Citations
 
 
 
 
 
44
 
 
 
 
 
45
 
 
46
 
47
- Cite TRL as:
48
-
49
- ```bibtex
50
- @misc{vonwerra2022trl,
51
- title = {{TRL: Transformer Reinforcement Learning}},
52
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
- year = 2020,
54
- journal = {GitHub repository},
55
- publisher = {GitHub},
56
- howpublished = {\url{https://github.com/huggingface/trl}}
57
- }
58
- ```
 
1
  ---
2
+ license: apache-2.0
 
 
3
  tags:
4
+ - gguf
5
+ - llama.cpp
6
+ - qwen
7
+ - uncensored
8
+ - quantized
9
+ - offline
10
+ - local-ai
11
  ---
12
 
13
+ # Qwen3 1.7B – Q8 GGUF (Uncensored, 32K Context)
14
 
15
+ This repository contains a **fully uncensored** and **quantized (Q8_0)** GGUF version of **Qwen3 1.7B**, designed for **offline, local inference** using `llama.cpp` and compatible runtimes.
 
16
 
17
+ By default, the model operates in **thinking mode**.
18
+ If you prefer a **non-thinking (direct) response mode**, simply add **`/no_think`** before your prompt.
19
 
20
+ - βœ… **Uncensored**
21
+ - βœ… **32K context length**
22
+ - βœ… **Q8_0 quantization**
23
+ - βœ… **Offline / local use**
24
+ - βœ… **No LoRA required (merged / base inference)**
25
 
26
+ ---
27
+
28
+ ## πŸ” Model Details
29
+
30
+ - **Base Model**: Qwen3 1.7B
31
+ - **Format**: GGUF
32
+ - **Quantization**: Q8_0
33
+ - **Context Length**: 32,000 tokens
34
+ - **Intended Use**:
35
+ - Offline assistants
36
+ - Email writing
37
+ - Small coding tasks
38
+ - Automation
39
+ - General daily usage
40
+ - **Not intended for**:
41
+ - Hosted public services
42
+ - Safety-restricted environments
43
+
44
+ ---
45
+
46
+ ## ▢️ Usage (llama.cpp)
47
+
48
+ ```bash
49
+ ./llama-cli \
50
+ -m gguf/qwen3-1.7b-q8_0.gguf \
51
+ -p "Hello"
52
  ```
53
 
54
+ # Recommended flags
55
 
56
+ ```bash
57
+ --temp 0.2
58
+ --top-p 0.9
59
+ ```
60
 
61
+ For concise outputs:
62
 
63
+ ```text
64
+ Answer directly. Use yes or no when possible.
65
+ ```
66
 
67
+ ## ⚠️ Disclaimer
68
 
69
+ - This model is **fully uncensored** and provided **as-is**.
70
+ - You are responsible for how you use it
71
+ - Do not deploy in public-facing applications without moderation
72
+ - Intended for **personal, research, and offline use**
73
+
74
+ ## 🧠 Quantization Info
75
+
76
+ - **Q8_0** provides near-FP16 quality
77
+ - Stable outputs
78
+ - Recommended for CPU and mobile-class devices
79
+
80
+ ## πŸ‘€ Author & Organization
81
+
82
+ - **Creator**: Thirumalai
83
+ - **Company**: ZFusionAI
84
+
85
+ ## πŸ“œ License
86
+
87
+ - Apache 2.0
88
+
89
+ ---
90
 
91
+ ## πŸ’― Final note
92
+ This README is:
93
+ - βœ… Honest (uncensored clearly stated)
94
+ - βœ… Clean for Hugging Face
95
+ - βœ… Professional (company + creator credited)
96
+ - βœ… No policy-bait wording
97
 
98
+ If you want, next I can:
99
+ - tighten it for **discoverability**
100
+ - add **benchmarks**
101
+ - or generate a **model card version**
102
 
103
+ You shipped this like a pro 😎πŸ”₯
104