GGUF
conversational
N8Programs commited on
Commit
0e20d07
·
verified ·
1 Parent(s): 4dc3d45

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +188 -3
README.md CHANGED
@@ -1,3 +1,188 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - N8Programs/unslop-good
5
+ base_model:
6
+ - Qwen/Qwen3-VL-30B-A3B-Instruct
7
+ ---
8
+ # Unslopper-30B-A3B
9
+
10
+ ## Model Description
11
+
12
+ **Unslopper-30B-A3B** is a fine-tuned language model designed to transform AI-generated text into more human-like prose while preserving semantic content. The model takes passages exhibiting typical AI writing patterns and rewrites them to sound more natural, varied, and authentic.
13
+
14
+ - **Base Model**: Qwen3-VL-Text-30B-A3B-Instruct (6-bit quantized)
15
+ - **Architecture**: Mixture of Experts (MoE) with 30B total parameters, 3B active
16
+ - **Fine-tuning Method**: LoRA (Low-Rank Adaptation)
17
+ - **Framework**: MLX (Apple Silicon optimized)
18
+
19
+ ## Intended Use
20
+
21
+ The model is intended to:
22
+ - Improve the naturalness of AI-generated creative writing
23
+ - Reduce detectable AI patterns in text (stylistic homogeneity, predictable phrasing)
24
+ - Serve as a post-processing step for AI writing assistants
25
+
26
+ **Not intended for**: Bypassing AI detection for academic dishonesty, fraud, or deceptive purposes.
27
+
28
+ ## Training Data
29
+
30
+ ### Data Generation Pipeline
31
+
32
+ The training data was synthetically generated using a novel "reverse distillation" approach:
33
+
34
+ 1. **Source**: Human-written literary passages extracted from a Sam Paech [preference dataset](https://huggingface.co/datasets/sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo) sourced from Project Gutenberg.
35
+ 2. **AI-ification Process**: Each human passage was iteratively rewritten 10 times by GPT-4o-mini, progressively amplifying AI-typical writing patterns
36
+ 3. **Pair Creation**: Final pairs consist of (AI-refined passage → original human passage)
37
+
38
+ This creates a supervised learning signal where the model learns to *reverse* the AI-ification process. The full dataset can be found at [N8Programs/unslop-good](https://huggingface.co/datasets/N8Programs/unslop-good).
39
+
40
+ ### Dataset Statistics
41
+
42
+ | Metric | Value |
43
+ |--------|-------|
44
+ | Training examples | 1,000 |
45
+ | Refinement iterations per passage | 10 |
46
+ | Total API calls for data generation | 10,000 |
47
+ | Source | Literary fiction passages |
48
+
49
+ ## Training Configuration
50
+
51
+ ### Model Architecture
52
+
53
+ | Parameter | Value |
54
+ |-----------|-------|
55
+ | Base model | Qwen3-VL-Text-30B-A3B-Instruct |
56
+ | Quantization | 6-bit |
57
+ | Total parameters | 30B |
58
+ | Active parameters | 3B (MoE) |
59
+
60
+ ### LoRA Configuration
61
+
62
+ | Parameter | Value |
63
+ |-----------|-------|
64
+ | Rank | 8 |
65
+ | Scale (alpha) | 20.0 |
66
+ | Dropout | 0.0 |
67
+ | Layers fine-tuned | 48 |
68
+ | Target modules | `self_attn.q_proj`, `self_attn.v_proj`, `self_attn.k_proj`, `self_attn.o_proj`, `mlp.gate_proj`, `mlp.switch_mlp.gate_proj`, `mlp.switch_mlp.up_proj`, `mlp.switch_mlp.down_proj` |
69
+
70
+ ### Training Hyperparameters
71
+
72
+ | Parameter | Value |
73
+ |-----------|-------|
74
+ | Optimizer | Adam |
75
+ | Learning rate | 1e-4 |
76
+ | LR schedule | Cosine decay with warmup |
77
+ | Warmup steps | 10 |
78
+ | Warmup init LR | 1e-5 |
79
+ | Final LR | 1e-5 |
80
+ | Batch size | 1 |
81
+ | Gradient accumulation | 1 |
82
+ | Training iterations | 1,000 |
83
+ | Max sequence length | 6,144 |
84
+ | Gradient checkpointing | Enabled |
85
+
86
+ ### Optimizer Configuration
87
+
88
+ ```yaml
89
+ optimizer: adam
90
+ betas: [0.9, 0.9999]
91
+ eps: 1e-6
92
+ bias_correction: true
93
+ ```
94
+
95
+ ## Inference
96
+
97
+ ### Recommended Settings
98
+
99
+ ```python
100
+ from mlx_lm import load, generate
101
+ from mlx_lm.sample_utils import make_sampler, make_logits_processors
102
+
103
+ model, tokenizer = load("Unslopper-30B-A3B-6bit")
104
+
105
+ def unslop(passage: str) -> str:
106
+ prompt = f"Rewrite this AI passage to sound more humanlike:\n{passage}"
107
+ messages = [{"role": "user", "content": prompt}]
108
+
109
+ output = generate(
110
+ model,
111
+ tokenizer,
112
+ tokenizer.apply_chat_template(messages, add_generation_prompt=True),
113
+ max_tokens=4096,
114
+ sampler=make_sampler(temp=0.8),
115
+ logits_processors=make_logits_processors(repetition_penalty=1.1),
116
+ )
117
+ return output.strip()
118
+ ```
119
+
120
+ ### Inference Parameters
121
+
122
+ | Parameter | Recommended Value |
123
+ |-----------|-------------------|
124
+ | Temperature | 0.8 |
125
+ | Repetition penalty | 1.1 |
126
+ | Max tokens | 4096 |
127
+
128
+ ## Evaluation
129
+
130
+ ### Methodology
131
+
132
+ The model was evaluated on 100 short stories (~800 words each) generated by GPT-5.2. Each story was processed through Unslopper, and both versions were evaluated on:
133
+
134
+ 1. **AI Detection**: Pangram API (measures "humanness" as 1 - AI fraction)
135
+ 2. **Writing Quality**: Claude Opus 4.5 scoring on coherence, style, and general quality (1-10 scale). Weakest-point Quality is the minimum of the three scores.
136
+ 3. **Control**: As a control, stories were also passed through Qwen3 VL 30B A3B without fine-tuning to assess the effect of base model capabilities, with the same prompting and sampling settings as Unslopper. Notably, no significant humanness improvement was observed in this control, though the same decrease in quality was noted. This indicates that the humanness gains are attributable to the fine-tuning process rather than inherent model capabilities.
137
+
138
+ ### Results
139
+
140
+ | Metric | GPT-5.2 (Original) | Unslopped | Control (GPT-5.2 + Qwen3 VL 30B A3B) | Delta (Unslopped - Original) |
141
+ |--------|-------------------|-----------|--------------------------------------|-----------------------------|
142
+ | Mean Humanness | 0.000 ± 0.000 | 0.481 ± 0.039 | 0.003 ± 0.003 | +0.481 ± 0.039 |
143
+ | Weakest-Point Quality | 8.60 ± 0.06 | 7.96 ± 0.10 | 7.82 ± 0.12 | -0.64 ± 0.08 |
144
+ | AI Detection Label | 100% AI | 30% AI, 45% Mixed, 25% Human | 99% AI, 1% Mixed, 0% Human | — |
145
+
146
+ ### Comparison to Baselines
147
+
148
+ | Model | Weakest-Point Quality (Mean) |
149
+ |-------|------------------------------|
150
+ | **Unslopped (GPT-5.2 + Unslopper)** | **7.96 ± 0.10** |
151
+ | Control (GPT-5.2 + Qwen3 VL 30B A3B) | 7.82 ± 0.12 |
152
+ | GPT-5.2 (Original) | 8.60 ± 0.06 |
153
+ | Mistral Large 3 (2512) | 6.64 ± 0.08 |
154
+ | GPT-4o Mini | 5.24 ± 0.06 |
155
+
156
+ ### Key Findings
157
+
158
+ 1. **Humanness significantly improves**: From 0.000 ± 0.000 to 0.481 ± 0.039 on the Pangram scale
159
+ 2. **Quality trade-off is modest**: 0.64 ± 0.08 point decrease in weakest-point score
160
+ 3. **Still competitive**: Unslopped output quality exceeds Mistral Large 3 and GPT-4o Mini baselines
161
+ 4. **AI detection effectiveness**: 70% of unslopped stories are no longer classified as pure "AI"
162
+
163
+ ## Limitations
164
+
165
+ - **Quality-humanness trade-off**: Some reduction in writing quality is expected
166
+ - **Domain specificity**: Trained primarily on literary fiction; may generalize less well to technical or academic writing
167
+ - **Detection arms race**: AI detection methods evolve; effectiveness may vary over time
168
+ - **Semantic drift**: Minor semantic changes may occur during rewriting
169
+
170
+ ## Ethical Considerations
171
+
172
+ This model demonstrates that AI-generated text can be made to appear more human-like. Users should:
173
+
174
+ - Use responsibly and transparently
175
+ - Not use for academic fraud or deceptive purposes
176
+ - Consider disclosure requirements in relevant contexts
177
+ - Be aware of potential misuse implications
178
+
179
+ ## Citation
180
+
181
+ ```bibtex
182
+ @misc{unslopper2025,
183
+ title={Unslopper-30B-A3B: Humanizing AI-Generated Text via Reverse Distillation},
184
+ author={N8Programs},
185
+ year={2025},
186
+ howpublished={LoRA fine-tune of Qwen3-VL-Text-30B-A3B-Instruct}
187
+ }
188
+ ```