Install from WinGet (Windows)
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0# Run inference directly in the terminal:
llama-cli -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0Use pre-built binary
# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0# Run inference directly in the terminal:
./llama-cli -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0Build from source code
git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0# Run inference directly in the terminal:
./build/bin/llama-cli -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0Use Docker
docker model run hf.co/WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0This is a duplicate of Cordux/flux2-klein-4B-uncensored-text-encoder for Flux-Klein-4B-CPU Space
Qwen3-4B Ablated (Uncensored) Text Encoder - GGUF Q4_0
Uncensored/ablated version of Qwen3-4B text encoder in GGUF Q4_0 format for Flux2 Klein 4B models.
Compatible Models
- Flux2 Klein 4B (Distilled & Base)
What This Does
This is an ablated (safety-filtering removed) text encoder that allows Flux2 Klein models to generate NSFW content without prompt censorship.
The base Qwen3-4B text encoder that ships with Flux2 Klein has safety filtering that prevents certain prompts from being processed properly.
Installation
- Download
qwen3-4b-abl-q4_0.gguf - Place in
ComfyUI/models/text_encoders/orComfyUI/models/unet/(for GGUF loaders) - In your workflow, use a GGUF-compatible text encoder loader node
- Point it to this file instead of the default Qwen3-4B
Prompting Tips
- Use "wearing nothing" instead of "naked/nude" for best nude results
- The model looks for clothing descriptors - even "nothing" counts as one
- Clinical terms like "vagina" don't work better than colloquial terms
- For explicit content beyond nudity, you'll need an NSFW LoRA
Language-Style Mapping Research
I discovered Flux.2 Klein associates languages with specific styles:
Japaneseโanime portraits, Germanโillustrated art, etc.
Full study here
Limitations
This removes prompt filtering but doesn't add visual knowledge. The base Flux2 Klein models have limited training on explicit content, so:
- โ Nudity works well
- โ Suggestive poses work
- โ Explicit anatomy requires a LoRA
- โ Sexual acts require a LoRA
Credits
- Based on huihui-ai/Qwen3-4B-abliterated
- Converted with llama.cpp
- Downloads last month
- 2,000
4-bit
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0# Run inference directly in the terminal: llama-cli -hf WeReCooking/flux2-klein-4B-uncensored-text-encoder:Q4_0