Refusal in Language Models Is Mediated by a Single Direction
Paper
•
2406.11717
•
Published
•
6
This is an uncensored version of xLAM created using Blasphemer.
| File | Size | Use Case |
|---|---|---|
| Q4_K_M | ~4.5GB | Best balance - most popular |
| Q5_K_M | ~5.5GB | Higher quality, slightly larger |
| F16 | ~15GB | Full precision (for further quantization) |
./llama-cli -m xLAM-f16.gguf -p "Your prompt here"
from llama_cpp import Llama
llm = Llama(
model_path="Llama-3.1-8B-Blasphemer-Q4_K_M.gguf",
n_ctx=8192,
n_gpu_layers=-1 # Use GPU
)
response = llm("Your prompt here", max_tokens=512)
print(response['choices'][0]['text'])
Abliteration removes refusal behavior from language models by identifying and removing the neural directions responsible for safety alignment. This is done through:
The result is a model that maintains capabilities while removing refusal behavior.
This model has reduced safety guardrails. Users are responsible for:
Compared to the original Llama:
If you use this model, please cite:
@software{blasphemer2024,
author = {Bradford, Christopher},
title = {Blasphemer: Abliteration for Language Models},
year = {2024},
url = {https://github.com/sunkencity999/blasphemer}
}
@article{arditi2024refusal,
title={Refusal in Language Models Is Mediated by a Single Direction},
author={Arditi, Andy and Obmann, Oscar and Syed, Aaquib and others},
journal={arXiv preprint arXiv:2406.11717},
year={2024}
}
This model inherits the Llama 3.1 license from Meta AI. Please review the Llama 3.1 License for usage terms.
4-bit
6-bit
8-bit
16-bit
Base model
Salesforce/Llama-xLAM-2-8b-fc-r