| --- |
| language: en |
| license: mit |
| library_name: transformers |
| pipeline_tag: text-generation |
| tags: |
| - text-generation |
| - ai-detection |
| - paraphrasing |
| - originality |
| - privacy |
| datasets: |
| - checkgpt |
| base_model: Qwen/Qwen2.5-3B-Instruct |
| model_type: causal-lm |
| --- |
| |
| # AuthorMist Originality |
|
|
| [](https://huggingface.co/authormist/originality) |
| [](https://opensource.org/licenses/MIT) |
|
|
| ## Overview |
|
|
| AuthorMist Originality is a specialized language model designed to transform AI-generated text into more human-like writing while preserving the original meaning. This model was developed using reinforcement learning techniques to specifically evade AI text detection systems, with a focus on Originality.ai's detection algorithms. |
|
|
| The model is based on Qwen2.5-3B Instruct and has been fine-tuned using Group Relative Policy Optimization (GRPO) with detector feedback as a reward signal. AuthorMist Originality demonstrates strong performance in reducing detectability across multiple AI text detection systems while maintaining high semantic similarity with the original text. |
|
|
| ## Key Features |
|
|
| - **Detector Evasion**: Trained specifically to evade Originality.ai's detection algorithms, with strong cross-detector generalization |
| - **Meaning Preservation**: Maintains high semantic similarity (>0.94) with the original text |
| - **Natural Output**: Produces fluent, coherent text that reads naturally |
| - **Broad Applicability**: Effective across various domains including academic, technical, and creative writing |
|
|
| ## Model Details |
|
|
| - **Base Model**: Qwen2.5-3B Instruct |
| - **Training Method**: Reinforcement Learning with Group Relative Policy Optimization (GRPO) |
| - **Training Data**: 10,000 human-written abstracts from the CheckGPT dataset with corresponding AI-generated versions |
| - **Domains Covered**: Computer Science, Humanities, Social Sciences, Physics, and more |
| - **Text Length Support**: Optimized for texts ranging from 100 to 500 words |
|
|
| ## Performance |
|
|
| AuthorMist Originality demonstrates exceptional performance in evading AI text detection: |
|
|
| - **Mean AUROC**: 0.49 across six major detection systems |
| - **Mean F1-score**: 0.09 across all tested detectors |
| - **Semantic Similarity**: >0.94 with original text |
|
|
| The model shows particularly strong performance against: |
| - Hello SimpleAI (AUROC: 0.07) |
| - Sapling (AUROC: 0.13) |
| - Winston.ai (AUROC: 0.35) |
|
|
| ## Usage |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| |
| # Load model and tokenizer |
| model_name = "authormist/authormist-originality" |
| tokenizer = AutoTokenizer.from_pretrained(model_name) |
| model = AutoModelForCausalLM.from_pretrained(model_name) |
| |
| # Prepare input text |
| ai_text = "Your AI-generated text here..." |
| prompt = f"""Please paraphrase the following text to make it more human-like while preserving the original meaning: |
| |
| {ai_text} |
| |
| Paraphrased text:""" |
| |
| # Generate paraphrased text |
| inputs = tokenizer(prompt, return_tensors="pt") |
| outputs = model.generate( |
| inputs.input_ids, |
| max_new_tokens=512, |
| temperature=0.7, |
| top_p=0.9, |
| do_sample=True |
| ) |
| paraphrased_text = tokenizer.decode(outputs[0], skip_special_tokens=True) |
| print(paraphrased_text.split("Paraphrased text:")[1].strip()) |
| ``` |
|
|
| ## Ethical Considerations |
|
|
| AuthorMist Originality is released for research purposes to advance understanding of AI text detection limitations and privacy-preserving technologies. We acknowledge the dual-use nature of this technology and emphasize the following ethical considerations: |
|
|
| 1. **Academic Integrity**: This model should not be used to misrepresent AI-generated content as human-written in academic settings where such distinctions are ethically relevant. |
|
|
| 2. **Transparency**: We encourage users to maintain transparency about the use of AI assistance in content creation, even when using privacy-enhancing tools like AuthorMist. |
|
|
| 3. **Privacy Protection**: The primary legitimate use case for this technology is protecting author privacy and preventing unfair discrimination against AI-assisted writing in contexts where such assistance is permissible. |
|
|
| 4. **Research Value**: This model provides valuable insights into the limitations of current AI detection systems and contributes to the ongoing research dialogue about AI text detection and privacy. |
|
|
| ## Citation |
|
|
| If you use AuthorMist Originality in your research, please cite our paper: |
|
|
| ```bibtex |
| @article{authormist2025, |
| title={AuthorMist: Evading AI Text Detectors with Reinforcement Learning}, |
| author={David, Isaac and Gervais, Arthur}, |
| journal={arXiv preprint}, |
| year={2025} |
| } |
| ``` |
|
|
| ## License |
|
|
| This model is released under the [MIT License](https://opensource.org/licenses/MIT). |
|
|
| ## Acknowledgments |
|
|
| We thank the developers of Qwen2.5 for the base model and the creators of the CheckGPT dataset for providing valuable training data. |