File size: 6,263 Bytes
2c914eb | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 | # SAL Visualizations
## Understanding the Plots
---
## Overview
These visualizations demonstrate SAL's core concepts using synthetic data. They are designed to illustrate principles, not report experimental results.
---
## Plot A: Gradient Preservation

### What It Shows
This plot compares gradient suppression (protection) between:
- **Base (Red):** Standard training with no communication
- **SAL (Cyan):** Training with Communication Layer
### Reading the Plot
- **X-axis:** Training steps (0-2000)
- **Y-axis:** Gradient suppression percentage (0-60%)
### Key Insight
**Base training:** Gradient suppression stays flat around 5-10%. Everything gets modified.
**SAL training:** Gradient suppression increases to ~45% as stable patterns are identified and protected.
### What This Means
SAL learns which parameters are stable and progressively protects them. This is "ask before updating" in action β SAL measures stability and reduces updates to stable parameters.
---
## Plot B: Stability Spectrum

### What It Shows
The distribution of parameters across three stability states:
- **Protected (Cyan):** Identity core β 12%
- **Neutral (Gray):** Adaptive zone β 71%
- **Volatile (Red):** Learning edge β 17%
### Reading the Plot
- **X-axis:** Stability categories
- **Y-axis:** Percentage of parameters
### Key Insight
A healthy model has:
- Small protected core (~12%) β fundamental learned patterns
- Large neutral zone (~71%) β flexible but careful
- Active learning edge (~17%) β where new knowledge enters
### What This Means
Not all parameters are equal. SAL identifies which parameters belong to which category and treats them accordingly. The identity core is protected, the learning edge is free to change, and the neutral zone adapts carefully.
---
## Plot C: Emergence Map

### What It Shows
A semantic field visualization with:
- **X-axis:** Coherence score (internal consistency)
- **Y-axis:** Novelty score (difference from known patterns)
- **Color:** Emergence intensity
- **Contours:** Density of states
### Reading the Plot
Each point is a semantic state. Clusters indicate natural organization:
- **Bottom-right:** High coherence, low novelty β Stable core
- **Top-left:** Low coherence, high novelty β Exploratory edge
- **Top-right:** High coherence, high novelty β **Emergent zone** (circled)
### Key Insight
**Emergence = Coherent Novelty**
True emergence requires BOTH:
- High coherence (structured, meaningful)
- High novelty (genuinely new)
Pure novelty without coherence = chaos.
Pure coherence without novelty = repetition.
### What This Means
SAL observes this field to detect emergence. When a pattern appears in the emergent zone (high coherence + high novelty), it represents genuine new learning that should be integrated and eventually protected.
---
## Plot D: Drift Reduction

### What It Shows
Semantic drift over training iterations:
- **Baseline (Red):** Exponential drift without protection
- **SAL (Cyan):** Stabilized drift with Communication Layer
- **Yellow dashed:** Critical drift threshold
### Reading the Plot
- **X-axis:** Training iterations (0-1000)
- **Y-axis:** Drift amount (0-1, where 1 = complete divergence)
### Key Insight
**Baseline:** Drift increases exponentially, crossing the critical threshold around iteration 400. This is catastrophic forgetting in action.
**SAL:** Drift stabilizes around 0.15-0.18, never approaching the critical threshold. **73% reduction in drift.**
### What This Means
Without protection, models gradually lose coherence as training overwrites stable patterns. SAL prevents this by protecting stable parameters, maintaining self-coherence throughout training.
---
## Plot E: Pulse-Split-Cascade Flow

### What It Shows
The PSC (Pulse-Split-Cascade) architecture:
```
PROMPT
β
Pulse 1 Pulse 2 Pulse 3 Pulse 4 Pulse 5 Pulse 6
β β β β β β
ββββββββββ΄βββββββββ ββββββββββ΄βββββββββ
β β
Lineage A β Lineage B
ββββββββββββ¬βββββββββββββββ
β
EMERGENCE
```
### Reading the Diagram
1. **Prompt** initiates the cascade
2. **Pulses** are independent semantic branches
3. **Lineages** form as pulses merge
4. **Selection** happens naturally (circled lineage = selected)
5. **Emergence** is the result
### Key Insight
**No rewards. No scores. Just resonance.**
Lineage A is selected not because it scored higher, but because it naturally became more coherent and resonant. This is semantic Game of Life β patterns emerge, compete, merge, and the most coherent persist.
### What This Means
PSC is SAL's approach to generation/inference. Instead of sampling and scoring, PSC:
- Generates multiple semantic branches
- Lets them evolve independently
- Observes which naturally become most coherent
- Selects through resonance, not ranking
---
## Generating These Plots
All plots are generated from synthetic SAL-conformant data. No real training data, human labels, or reward signals are used.
### Running the Scripts
```bash
cd scripts/
python plot_A_gradient_preservation.py
python plot_B_stability_spectrum.py
python plot_C_emergence_map.py
python plot_D_drift_reduction.py
python plot_E_psc_flow.py
```
### Requirements
```
numpy
matplotlib
scipy
```
---
## Terminology Note
These plots deliberately avoid RLHF/Safety/Reward terminology:
| β Avoided | β
Used |
|-----------|--------|
| reward | coherence_score |
| loss | drift_amount |
| policy | lineage |
| human feedback | stability_measurement |
| alignment | coherence |
This is intentional. SAL is a different paradigm β the language reflects that.
---
*For technical details, see [Architecture](architecture.md).*
*For philosophy, see [Principles](principles.md).*
|