AI & ML interests

RAG + Fine-Tuning.

Recent Activity

Organization Card

Divinci AI

Feature-level interpretability artifacts for open transformers — built openly, validated empirically.

A vindex is a transformer's weights decompiled into a queryable feature database. It exposes the entity associations, circuit structure, and knowledge-editing surfaces that live inside a model's FFN layers — without requiring GPU inference for most operations.

Think of it as the model's index: the thing you search before you run it.


Interactive viewer

LarQL Vindex Viewer — interactive 3D + 2D circuit visualization

→ Open the interactive viewer

Pick any of 9 models from the dropdown. Toggle between the 3D cylinder spiral and a flat 2D circuit/network view. Hit ⇌ Compare to render the current model alongside Bonsai 1-bit, side-by-side — the contrast between fp16 structure (organized rings) and 1-bit dissolution (scattered cloud) is the most direct picture of what 1-bit training does to a transformer's internal organization that we know how to render. Search for entity features (?q=paris&model=gemma-4-e2b) to see real probe-derived activations light up across the layer stack — backed by a 5000-token offline-built search index.


Published vindexes

Cross-family evidence in hand: Gemma, Qwen3, Mistral, Llama, OpenAI MoE, Moonshot MoE, DeepSeek-V4 MoE, plus two 1-bit controls.

MODELARCHITECTUREPARAMSVINDEXC4 (LAYER TEMP)NOTES
Gemma 4 E2B-itDense (Gemma 4)2Bgemma-4-e2b-vindex0.0407 ± 0.00043-seed validated; headline universal-constant model
Qwen3-0.6BDense (Qwen 3)0.6Bqwen3-0.6b-vindex0.411Smallest published; Qwen3 family-elevated C4
Qwen3-8B bf16Dense (Qwen 3)8Bqwen3-8b-vindex0.804Architecture control for Bonsai
Qwen3.6-35B-A3BMoE (Qwen 3.6)35B / 3B activeqwen3.6-35b-a3b-vindex256 experts, 40 layers
Ministral-3BDense (Mistral 3)3Bministral-3b-vindex0.265fp8 → bf16 reconstruction
Llama 3.1-8BDense (Llama 3.1)8Bllama-3.1-8b-vindex0.012Llama family signature
MedGemma 1.5-4BDense (Gemma multimodal)4Bmedgemma-1.5-4b-vindex1.898 ⚠45× cohort anomaly — under investigation
GPT-OSS 120BMoE (OpenAI)120Bgpt-oss-120b-vindexS[0] grows 117× with depth (L0=111 → final=13,056)
Kimi-K2-InstructMoE fp8-native (DeepSeek-V3 style)1T / 32B activekimi-k2-instruct-vindex0.0938 (MoE median)60 MoE layers; 42.28 GB gate_proj binary; broader L52–L60 secondary rise than initial dome SVD suggested
DeepSeek-V4-FlashMoE MXFP4 (DeepSeek-V4)43L / 256 experts / 6 activedeepseek-v4-flash-vindex0.108 (MoE median)43-layer all-MoE; 11.54 GB gate_proj binary; first-peak L18 + double-bend profile (distinct from Kimi smooth dome); MXFP4 expert unpacking
DeepSeek-V4-ProMoE MXFP4 (DeepSeek-V4)61L / 384 experts / 6 activedeepseek-v4-pro-vindex0.0653 (MoE median)61-layer all-MoE; 42.98 GB gate_proj binary; lowest var@64 of 3 published MoE vindexes (V4-Pro 0.065 < Kimi 0.094 < V4-Flash 0.108) — V4-Pro experts are most shared/redundant; late secondary rise L53–L60
Bonsai 8B1-bit (Qwen 3 base, post-quantized)8Bvindex pending publish0.429C5 = 1 (circuit dissolved); var@64 = 0.093
BitNet b1.58-2B-4T1-bit (Microsoft, native)2Bvindex pending publish(Phase 2 pending)var@64 = 0.111 mean across 30 layers — n=2 confirmation of dissolution

What's a vindex?

Standard model weights tell you what a model computes. A vindex tells you where it stores specific knowledge and which features need to change for a targeted edit.

Concretely: given a query like "Paris → capital", a vindex walk returns the layers, feature directions, and token associations that encode that fact. A patch operation writes a rank-1 ΔW that suppresses or overwrites that association — compiled back to standard HuggingFace safetensors for inference.

LarQL (the toolchain that builds vindexes) is open-source: github.com/chrishayuk/larql | github.com/Divinci-AI/larql.


Research

Paper 1 — Architectural Invariants of Transformer Computation

arXiv preprint forthcoming

Five properties measured across every model in this collection. Three hold within ±15% coefficient of variation across architectures, organizations, and scales. One collapses under 1-bit quantization — replicated across two independent 1-bit models from two organizations (n = 2). One scales monotonically with model size.

The headline universal constant — layer temperature C4 — is reproducible at the 1% precision level: a three-seed run on Gemma 4 E2B gives C4 = 0.0407 ± 0.0004, with circuit-stage count perfectly stable (C5 = 4 ± 0) across all seeds.

Paper 2 — Constellation Edits

draft, arXiv after 3-seed runs + α-sweep appendix

Mechanistic knowledge editing in transformer feature space. Includes a negative result: why activation-space edits fail in 1-bit models, and what weight-space geometry reveals about why.

Companion blog series — The Interpretability Diaries

Working notebooks: github.com/Divinci-AI/server/tree/preview/notebooks


Working in public

Every measurement in our papers traces back to a notebook and a commit. Negative results ship alongside positive ones — the MLP compensation mechanism that defeats knowledge editing in 1-bit models is in the notebooks, not buried in a supplement.

If you replicate a result and find a discrepancy, open an issue on the LarQL repo.


Vindexes on this org are free for academic and research use (CC-BY-NC 4.0). Commercial licensing: mike@divinci.ai