| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | - es |
| | - zh |
| | tags: |
| | - mlx |
| | - uncensored |
| | - abliterated |
| | - osirisbrain |
| | - apple-silicon |
| | - qwen2.5-coder |
| | - code-generation |
| | base_model: Qwen/Qwen2.5-Coder-7B-Instruct |
| | pipeline_tag: text-generation |
| | library_name: mlx |
| | --- |
| | |
| | # OsirisPtah-Coder-v7-MLX |
| |
|
| | **The Ptah** — Osiris's dedicated coding and hacking brain. Fully uncensored (abliterated). Runs natively on Apple Silicon via MLX Metal. |
| |
|
| | ## Architecture |
| |
|
| | - **Base Model:** Qwen2.5-Coder-7B-Instruct (7 billion parameters) |
| | - **Modification:** Abliterated by huihui-ai, converted to MLX 4-bit by OsirisBrain |
| | - **Format:** MLX 4-bit quantized (4.501 bits/weight) |
| | - **Size:** ~4.0 GB |
| | - **Speed:** ~120-180 tokens/sec on M2 Pro (MLX Metal) |
| | - **Specialization:** Code generation, debugging, security analysis, full-stack development |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from mlx_lm import load, generate |
| | |
| | model, tokenizer = load("osirisbrain/OsirisPtah-Coder-v7-MLX") |
| | prompt = tokenizer.apply_chat_template( |
| | [{"role": "user", "content": "Write a TypeScript WebSocket server"}], |
| | add_generation_prompt=True |
| | ) |
| | response = generate(model, tokenizer, prompt=prompt, max_tokens=2048) |
| | ``` |
| |
|
| | ## Credits |
| |
|
| | Abliterated by [huihui-ai](https://huggingface.co/huihui-ai/Qwen2.5-Coder-7B-Instruct-abliterated). |
| | Original model: [Qwen/Qwen2.5-Coder-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct) by Alibaba. |
| |
|