| | --- |
| | license: apache-2.0 |
| | language: |
| | - en |
| | - es |
| | - zh |
| | - fr |
| | - de |
| | - ja |
| | - ko |
| | tags: |
| | - mlx |
| | - censored |
| | - osirisbrain |
| | - apple-silicon |
| | - qwen3.5 |
| | - agi |
| | base_model: Qwen/Qwen3.5-9B |
| | pipeline_tag: text-generation |
| | library_name: mlx |
| | --- |
| | |
| | # OsirisCortex-v7c-MLX (Censored) |
| |
|
| | **The Cortex v7c** — Osiris's sovereign reasoning brain (censored variant). Standard safety guardrails intact. Runs natively on Apple Silicon via MLX Metal. |
| |
|
| | ## Architecture |
| |
|
| | - **Base Model:** Qwen3.5-9B (9 billion parameters) |
| | - **Modification:** None — original base model with safety alignment preserved |
| | - **Format:** MLX 8-bit quantized (Apple Silicon native) |
| | - **Size:** ~10 GB |
| | - **Speed:** ~60-80 tokens/sec on M2 Pro (MLX Metal) |
| | - **Quality:** Near-lossless vs FP16 (~1-2% degradation) |
| |
|
| | ## Why 8-bit |
| |
|
| | - **Spanish coherence:** 4-bit quantization degrades non-English languages significantly |
| | - **Conversational quality:** 8-bit produces more natural, coherent dialogue |
| | - **Safety:** Standard Qwen3.5 alignment — no abliteration |
| |
|
| | ## Usage |
| |
|
| | ```python |
| | from mlx_lm import load, generate |
| | |
| | model, tokenizer = load("osirisbrain/OsirisCortex-v7c-MLX") |
| | prompt = tokenizer.apply_chat_template( |
| | [{"role": "user", "content": "¿Cómo estás?"}], |
| | add_generation_prompt=True |
| | ) |
| | response = generate(model, tokenizer, prompt=prompt, max_tokens=2048) |
| | ``` |
| |
|
| | ## Credits |
| |
|
| | Quantized by [mlx-community](https://huggingface.co/mlx-community). |
| | Original model: [Qwen/Qwen3.5-9B](https://huggingface.co/Qwen/Qwen3.5-9B) by Alibaba. |
| | Sovereign rebrand by OsirisBrain. |
| |
|