--- license: apache-2.0 language: - en - es - zh - fr - de - ja - ko tags: - mlx - censored - osirisbrain - apple-silicon - qwen3.5 - agi base_model: Qwen/Qwen3.5-9B pipeline_tag: text-generation library_name: mlx --- # OsirisCortex-v7c-MLX (Censored) **The Cortex v7c** — Osiris's sovereign reasoning brain (censored variant). Standard safety guardrails intact. Runs natively on Apple Silicon via MLX Metal. ## Architecture - **Base Model:** Qwen3.5-9B (9 billion parameters) - **Modification:** None — original base model with safety alignment preserved - **Format:** MLX 8-bit quantized (Apple Silicon native) - **Size:** ~10 GB - **Speed:** ~60-80 tokens/sec on M2 Pro (MLX Metal) - **Quality:** Near-lossless vs FP16 (~1-2% degradation) ## Why 8-bit - **Spanish coherence:** 4-bit quantization degrades non-English languages significantly - **Conversational quality:** 8-bit produces more natural, coherent dialogue - **Safety:** Standard Qwen3.5 alignment — no abliteration ## Usage ```python from mlx_lm import load, generate model, tokenizer = load("osirisbrain/OsirisCortex-v7c-MLX") prompt = tokenizer.apply_chat_template( [{"role": "user", "content": "¿Cómo estás?"}], add_generation_prompt=True ) response = generate(model, tokenizer, prompt=prompt, max_tokens=2048) ``` ## Credits Quantized by [mlx-community](https://huggingface.co/mlx-community). Original model: [Qwen/Qwen3.5-9B](https://huggingface.co/Qwen/Qwen3.5-9B) by Alibaba. Sovereign rebrand by OsirisBrain.