sal-learning / docs /plots.md
Whiteroom
Initial SAL core for HF (no plots/pdf)
2c914eb

SAL Visualizations

Understanding the Plots


Overview

These visualizations demonstrate SAL's core concepts using synthetic data. They are designed to illustrate principles, not report experimental results.


Plot A: Gradient Preservation

Gradient Preservation

What It Shows

This plot compares gradient suppression (protection) between:

  • Base (Red): Standard training with no communication
  • SAL (Cyan): Training with Communication Layer

Reading the Plot

  • X-axis: Training steps (0-2000)
  • Y-axis: Gradient suppression percentage (0-60%)

Key Insight

Base training: Gradient suppression stays flat around 5-10%. Everything gets modified.

SAL training: Gradient suppression increases to ~45% as stable patterns are identified and protected.

What This Means

SAL learns which parameters are stable and progressively protects them. This is "ask before updating" in action β€” SAL measures stability and reduces updates to stable parameters.


Plot B: Stability Spectrum

Stability Spectrum

What It Shows

The distribution of parameters across three stability states:

  • Protected (Cyan): Identity core β€” 12%
  • Neutral (Gray): Adaptive zone β€” 71%
  • Volatile (Red): Learning edge β€” 17%

Reading the Plot

  • X-axis: Stability categories
  • Y-axis: Percentage of parameters

Key Insight

A healthy model has:

  • Small protected core (~12%) β€” fundamental learned patterns
  • Large neutral zone (~71%) β€” flexible but careful
  • Active learning edge (~17%) β€” where new knowledge enters

What This Means

Not all parameters are equal. SAL identifies which parameters belong to which category and treats them accordingly. The identity core is protected, the learning edge is free to change, and the neutral zone adapts carefully.


Plot C: Emergence Map

Emergence Map

What It Shows

A semantic field visualization with:

  • X-axis: Coherence score (internal consistency)
  • Y-axis: Novelty score (difference from known patterns)
  • Color: Emergence intensity
  • Contours: Density of states

Reading the Plot

Each point is a semantic state. Clusters indicate natural organization:

  • Bottom-right: High coherence, low novelty β†’ Stable core
  • Top-left: Low coherence, high novelty β†’ Exploratory edge
  • Top-right: High coherence, high novelty β†’ Emergent zone (circled)

Key Insight

Emergence = Coherent Novelty

True emergence requires BOTH:

  • High coherence (structured, meaningful)
  • High novelty (genuinely new)

Pure novelty without coherence = chaos. Pure coherence without novelty = repetition.

What This Means

SAL observes this field to detect emergence. When a pattern appears in the emergent zone (high coherence + high novelty), it represents genuine new learning that should be integrated and eventually protected.


Plot D: Drift Reduction

Drift Reduction

What It Shows

Semantic drift over training iterations:

  • Baseline (Red): Exponential drift without protection
  • SAL (Cyan): Stabilized drift with Communication Layer
  • Yellow dashed: Critical drift threshold

Reading the Plot

  • X-axis: Training iterations (0-1000)
  • Y-axis: Drift amount (0-1, where 1 = complete divergence)

Key Insight

Baseline: Drift increases exponentially, crossing the critical threshold around iteration 400. This is catastrophic forgetting in action.

SAL: Drift stabilizes around 0.15-0.18, never approaching the critical threshold. 73% reduction in drift.

What This Means

Without protection, models gradually lose coherence as training overwrites stable patterns. SAL prevents this by protecting stable parameters, maintaining self-coherence throughout training.


Plot E: Pulse-Split-Cascade Flow

PSC Flow

What It Shows

The PSC (Pulse-Split-Cascade) architecture:

PROMPT
   ↓
Pulse 1  Pulse 2  Pulse 3  Pulse 4  Pulse 5  Pulse 6
   ↓        ↓        ↓        ↓        ↓        ↓
   β””β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜        β””β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”˜
           ↓                          ↓
      Lineage A βœ“                Lineage B
           β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
                      ↓
                 EMERGENCE

Reading the Diagram

  1. Prompt initiates the cascade
  2. Pulses are independent semantic branches
  3. Lineages form as pulses merge
  4. Selection happens naturally (circled lineage = selected)
  5. Emergence is the result

Key Insight

No rewards. No scores. Just resonance.

Lineage A is selected not because it scored higher, but because it naturally became more coherent and resonant. This is semantic Game of Life β€” patterns emerge, compete, merge, and the most coherent persist.

What This Means

PSC is SAL's approach to generation/inference. Instead of sampling and scoring, PSC:

  • Generates multiple semantic branches
  • Lets them evolve independently
  • Observes which naturally become most coherent
  • Selects through resonance, not ranking

Generating These Plots

All plots are generated from synthetic SAL-conformant data. No real training data, human labels, or reward signals are used.

Running the Scripts

cd scripts/
python plot_A_gradient_preservation.py
python plot_B_stability_spectrum.py
python plot_C_emergence_map.py
python plot_D_drift_reduction.py
python plot_E_psc_flow.py

Requirements

numpy
matplotlib
scipy

Terminology Note

These plots deliberately avoid RLHF/Safety/Reward terminology:

❌ Avoided βœ… Used
reward coherence_score
loss drift_amount
policy lineage
human feedback stability_measurement
alignment coherence

This is intentional. SAL is a different paradigm β€” the language reflects that.


For technical details, see Architecture. For philosophy, see Principles.