File size: 3,356 Bytes
4a98e60
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cb6f945
4a98e60
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cb6f945
 
4a98e60
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
# Exorobourii Research Lab

> **"What we observe is not nature itself, but nature exposed to our method of questioning."** — Werner Heisenberg

Exorobourii is a research initiative dedicated to **Mechanistic Interpretability** and **Efficient Intelligence**. We believe technology should be a glass box, not a black box. We build instruments to measure the internal physics of AI, and engineering frameworks to optimize it for ethical and ecological sustainability.

## 📡 The Mission

Current AI development faces an "Observability Crisis". We are building engines that are faster and more powerful, but we rely on a dashboard that only has a speedometer (`val_loss`).

Our work focuses on three pillars:
1.  **Observability:** Developing the **VSM Protocol** to act as a "mechanistic stethoscope" for Transformer attention.
2.  **Efficiency:** Engineering **Nano-LLMs** (Project Janus) that achieve "Super-Chinchilla" performance by eliminating structural redundancy.
3.  **Sustainability:** Reducing the computational cost of intelligence through **Vector Space Homeostasis**.

---

## 🔬 Key Research Initiatives

### 1. The VSM Protocol
*A Framework for Quantifying and Guiding Attention Head Specialization.*

The VSM Protocol treats the Transformer architecture as a physical system for spectral processing. We utilize two novel metrics to track the evolution of a model's "mind" from initialization to convergence:

* **$\sigma_p$ (Coherence):** Measures the focus/entropy of attention heads.
* **$\sigma_a$ (Novelty/Agreement):** Measures the degree of cross-head specialization.

Our research has quantified the "Untrained Symmetry" phenomenon (Softmax Collapse) and mapped the "Diagonally Oppositional" trajectory of healthy learning.

### 2. Project Janus
*Engineering Efficient Nano-LLMs via Feature Orthogonality.*

Project Janus is an attempt to solve "Attentional Collapse"—the tendency for small models (Nano-LLMs) to learn redundant features due to limited capacity.

By implementing **Vector Space Homeostasis** (a diversity pressure term $\lambda_{div}$ in the loss function) and a **Trapezoidal Pressure Schedule**, we force the model to maintain feature orthogonality.

**Key Results (Janus v3 vs. Baseline):**
* **Architecture:** 40M Parameters (Llama-style Chassis).
* **Performance:** 9.2% reduction in Loss on logical coherence tasks.
* **Efficiency:** Achieved parity loss with **28% less structural redundancy**.
* **Generalization:** 0.91 improvement in Perplexity on WikiText-103.

![Efficiency Gap Chart - Janus Loss vs Baseline](path/to/efficiency_gap_chart.png)

---

## 🛠️ Usage & Citation

We believe in open science. Our protocols and model weights are released here to encourage the community to move beyond black-box optimization.

### BibTeX

If you use the VSM Protocol or Janus methodology in your research, please cite:

```bibtex
@techreport{belanger2025vsm,
  title={The VSM (Vector-Space-Mapping) Protocol: A Framework for Quantifying and Guiding Attention Head Specialization in Transformers},
  author={Belanger, Jonathan R.},
  institution={Exorobourii},
  year={2025}
}

@techreport{belanger2025janus,
  title={Project Janus: Engineering Efficient Nano-LLMs via Feature Orthogonality and Vector Space Homeostasis},
  author={Belanger, Jonathan R.},
  institution={Exorobourii},
  year={2025}
}