HereticFT-Antislop

HereticFT-Antislop is a refined version of DrRiceIO7/HereticFT, a Gemma-3 4B based model. This version has been specifically fine-tuned to eliminate common "AI slop"β€”over-represented words, phrases, and repetitive n-gramsβ€”using the Auto-Antislop pipeline.

πŸš€ Overview

The goal of this model is to maintain the creative, uncensored and unique personality of the base model while stripping away the predictable linguistic patterns often found in modern LLMs (e.g., "tapestry," "testament," "delve," "it's important to remember").

πŸ› οΈ How it was made

This model was created using the Auto-Antislop pipeline developed by Sam Paech.

The Process:

  1. Slop Identification: The base model was analyzed on a large set of creative writing prompts to identify its unique "slop profile"β€”the words and phrases it over-uses compared to human writing.
  2. Preference Dataset Generation: Using antislop-vllm, a preference dataset was generated. When the model attempted to use "slop" tokens, the sampler diverted it to more coherent, human-like alternatives.
  3. FTPO Fine-tuning: The model underwent Final-Token Preference Optimisation (FTPO). Unlike standard DPO, FTPO is a surgical fine-tuning method that specifically targets the logits of the "slop" tokens and their preferred alternatives, minimizing general model degradation and preserving the original model's strengths.

πŸ“ˆ Improvements

  • Reduced Repetition: Lowered frequency of over-represented n-grams and common AI clichΓ©s.
  • Enhanced Vocabulary: Encourages more diverse and human-like word choices.
  • Preserved Personality: The "Heretic" edge remains intact, but the prose is cleaner and more professional.

πŸ§ͺ Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "DrRiceIO7/HereticFT-Antislop"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype="auto", device_map="auto")

prompt = "Write a short story about a heretic in a high-tech dystopia."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

🀝 Acknowledgments


Disclaimer: This model description was generated by Gemini 3 Flash Preview.

Downloads last month
1
Safetensors
Model size
4B params
Tensor type
F32
Β·
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for DrRiceIO7/HereticFT-Antislop

Finetuned
(2)
this model