| --- |
| base_model: |
| - huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated |
| datasets: |
| - Aquiles-ai/Athenea-Coding-100k |
| license: apache-2.0 |
| tags: |
| - code |
| - agent |
| - text-generation-inference |
| - merge |
| - uncensored |
| - athenea |
| language: |
| - en |
| - es |
| pipeline_tag: text-generation |
| library_name: transformers |
| --- |
| |
| <h1 align="center">Athenea-4B-Coding</h1> |
|
|
|  |
|
|
| **Athenea-4B-Coding** is a fine-tuned version of [huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated](https://huggingface.co/huihui-ai/Huihui-Qwen3-4B-Thinking-2507-abliterated), specialized in **code reasoning, debugging, and problem solving**. |
| Trained on high-quality programming data with explicit reasoning traces using `<think>` and `</think>` tags, the model is designed to perform detailed step-by-step reasoning for software development, algorithm design, and code comprehension tasks. |
|
|
| > ⚠️ **Important Note:** This model uses an *abliterated (uncensored)* base version, providing full expressive freedom and unrestricted output generation. Users are fully responsible for any use or content produced by the model. It is intended exclusively for research and experimentation purposes. |
|
|
| ## 🎯 Model Description |
|
|
| Athenea-4B-Coding extends Huihui-Qwen3’s structured reasoning capabilities into programming-related domains, showing strong performance on logical problem-solving, code completion, and debugging scenarios. |
|
|
| Key features: |
|
|
| * **Step-by-step code reasoning** within `<think>` blocks |
| * **Specialization in algorithmic and debugging tasks** |
| * **Uncensored output generation** for full reasoning visibility |
| * **Improved logical consistency** through focused fine-tuning |
| * **Compatible with open inference frameworks** (Transformers, vLLM, etc.) |
|
|
| The model was fine-tuned using the dataset [Aquiles-ai/Athenea-Coding-100k](https://huggingface.co/datasets/Aquiles-ai/Athenea-Coding-100k), which includes diverse programming challenges, structured reasoning chains, and natural language explanations across multiple programming languages. |
|
|
| > Note: Fine-tuning was performed using **Kronos**, Aquiles-ai’s proprietary enterprise fine-tuning system. |
|
|
| ## 💻 Usage |
|
|
| ### Installation |
|
|
| ```bash |
| uv pip install transformers torch accelerate |
| ``` |
|
|
| ### Basic Inference |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| import torch |
| |
| model = AutoModelForCausalLM.from_pretrained("Aquiles-ai/Athenea-4B-Coding", |
| dtype=torch.bfloat16, |
| trust_remote_code=True, |
| device_map="auto", |
| attn_implementation="flash_attention_2") # Requires flash-attn |
| |
| # Without flash-attn: |
| # model = AutoModelForCausalLM.from_pretrained("Aquiles-ai/Athenea-4B-Coding", |
| # dtype="auto", |
| # device_map="auto" |
| # ) |
| |
| tokenizer = AutoTokenizer.from_pretrained("Aquiles-ai/Athenea-4B-Coding", trust_remote_code=True) |
| |
| messages = [ |
| {"role": "user", "content": "Hey, write a Python function that calculates the factorial of a number recursively."} |
| ] |
| |
| inputs = tokenizer.apply_chat_template( |
| messages, |
| add_generation_prompt=True, |
| tokenize=True, |
| return_dict=True, |
| return_tensors="pt", |
| ).to('cuda') |
| |
| with torch.no_grad(): |
| output = model.generate( |
| **inputs, |
| max_new_tokens=8092, |
| pad_token_id=tokenizer.eos_token_id, |
| eos_token_id=tokenizer.eos_token_id, |
| ) |
| |
| # Decode and print the output |
| print(tokenizer.decode(output[0], skip_special_tokens=True)) |
| ``` |
|
|
| ### Streaming Inference |
|
|
| ```python |
| from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer |
| import torch |
| from threading import Thread |
| |
| model = AutoModelForCausalLM.from_pretrained("Aquiles-ai/Athenea-4B-Coding", |
| dtype=torch.bfloat16, |
| trust_remote_code=True, |
| device_map="auto", |
| attn_implementation="flash_attention_2") |
| |
| tokenizer = AutoTokenizer.from_pretrained("Aquiles-ai/Athenea-4B-Coding", trust_remote_code=True) |
| |
| messages = [ |
| {"role": "user", "content": "Hey, write a Python function that implements the binary search algorithm recursively."} |
| ] |
| |
| inputs = tokenizer.apply_chat_template( |
| messages, |
| add_generation_prompt=True, |
| tokenize=True, |
| return_dict=True, |
| return_tensors="pt", |
| ).to('cuda') |
| |
| # Create the streamer |
| streamer = TextIteratorStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True) |
| |
| # Build kwargs for generate |
| generate_kwargs = dict( |
| **inputs, |
| max_new_tokens=8092, |
| pad_token_id=tokenizer.eos_token_id, |
| eos_token_id=tokenizer.eos_token_id, |
| streamer=streamer, |
| ) |
| |
| def _generate_thread(model, kwargs): |
| with torch.no_grad(): |
| model.generate(**kwargs) |
| |
| thread = Thread(target=_generate_thread, args=(model, generate_kwargs)) |
| thread.start() |
| |
| for chunk in streamer: |
| print(chunk, end="", flush=True) |
| ``` |
|
|
| ### Production Deployment with vLLM |
|
|
| **Start server:** |
|
|
| ```bash |
| vllm serve Aquiles-ai/Athenea-4B-Coding \ |
| --host 0.0.0.0 \ |
| --port 8000 \ |
| --api-key dummyapikey \ |
| --max-model-len=16384 \ |
| --async-scheduling \ |
| --gpu-memory-utilization=0.90 |
| ``` |
|
|
| **Request to the server from the OpenAI client:** |
|
|
| ```python |
| from openai import OpenAI |
| client = OpenAI(api_key="dummyapikey", base_url="http://127.0.0.1:8000/v1") |
| stream = client.chat.completions.create( |
| model="Aquiles-ai/Athenea-4B-Coding", |
| messages=[{ |
| "role": "user", |
| "content": "Hey, write a Python function that determines if a string is a palindrome, ignoring case, spaces, and punctuation." |
| }], |
| max_tokens=8092, |
| stream=True |
| ) |
| for chunk in stream: |
| if chunk.choices[0].delta.content: |
| print(chunk.choices[0].delta.content, end="", flush=True) |
| ``` |
|
|
| **vLLM Benefits:** 20-30x faster inference, OpenAI-compatible API, continuous batching, async scheduling. |
|
|
| ## Contact |
|
|
| - **More about [Aquiles-ai](https://aquiles-ai.vercel.app)).** |
|
|
| - **Aquiles-ai on [GitHub](https://github.com/Aquiles-ai).** |
|
|
| - **Our collections at [HuggingFace](https://huggingface.co/Aquiles-ai/collections).** |
|
|
| ### Aquiles-playground |
|
|
| In addition to code usage, you can also try our models locally through an [open-source playground on GitHub](https://github.com/Aquiles-ai/aquiles-playground). |
|
|
|  |
|
|
| <p align="center"> |
| Made with ❤️ by <a href="https://github.com/Aquiles-ai">Aquiles-ai</a> |
| </p> |