File size: 4,247 Bytes
1a60fa2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
---
license: apache-2.0
tags:
  - flow-matching
  - diffusion
  - geometric-deep-learning
  - constellation
  - geolip
  - cifar10
  - geometric-lookup
---

# GeoLIP Spherical Diffusion Prototype

**Flow matching diffusion through constellation bottleneck on S^15.**

Four progressive experiments proving that geometric triangulation
on the unit hypersphere is a viable information bottleneck for
diffusion models β€” and that the binding constant 0.29154 emerges
from velocity matching through geometric lookup.

## Experiments

### v1 β€” Regulator (baseline)
Constellation as a side-channel regulator on feature maps.
Gate stayed at 6%. Constellation was decorative.
- Loss: 0.1900 | Params: 6.1M | Near 0.29: 0%

### v2 β€” Skip Bypass (the sneaky test)
268M parameter `Linear(16384, 16384)` skip projection alongside
the constellation bottleneck. The model was given every reason
to bypass the constellation. **It chose the constellation** β€” gate
at 11.8%, routing 88% through 768 triangulation dimensions.
- Loss: 0.1757 | Params: 287M | Near 0.29: 9%

### v3 β€” Pure Constellation Bottleneck
Skip projection removed. Everything through S^15. Zero bypass.
Beat the 268M skip version with 8Γ— fewer bottleneck params.
Reconstruction cos_sim β‰ˆ 0 β€” the bottleneck is a geometric
lookup table, not an autoencoder.
- Loss: 0.1749 | Params: 36.6M | Near 0.29: 30%

### v4 β€” Geometric Lookup Flow Matching (GLFM)
Three-stage pipeline: Address β†’ Condition β†’ Generate.
Multi-scale addressing (coarse + fine). 46% of anchors
converged within Β±0.05 of the binding constant 0.29154.
- Loss: 0.1754 | Params: 35.2M | Near 0.29: 46%

## The 0.29154 Binding Constant

Anchor drift from home position converges toward 0.29154 radians
across all experiments. This constant has now appeared in:

| Domain | Architecture | Training |
|---|---|---|
| MinimalShunts | Binding/separation phase boundary | Contrastive |
| CLIP projections | Geometric transition | Contrastive |
| T5 generation | Alpha convergence | Language modeling |
| CaptionBERT | Phase boundary | Contrastive |
| **Flow matching** | **Max anchor drift** | **Velocity matching** |

The constant marks the boundary where anchors transition from
geometric frame holders to task-specific encoders.

## Key Empirical Results

| Finding | Result |
|---|---|
| CV β‰ˆ 0.20 is geometry of S^15 | Precision-invariant, 1-bit to fp64 |
| Constellation relay preserves 99.4% cos_to_orig at depth 16 | vs 7.4% for attention |
| Model prefers constellation over 268M skip bypass | 88/12 split |
| 768 tri dims match 16384 unconstrained dims for velocity | cos 0.949 |
| Bottleneck doesn't reconstruct β€” it's a lookup table | cos_sim β‰ˆ 0 to input |
| Anchors self-organize: structural (<0.29) vs semantic (>0.29) | Confirmed across 4 versions |

## Architecture β€” GLFM (v4)

```
Stage 1 β€” ADDRESS
  encoder(x_t) β†’ (B, 256, 8, 8)
  coarse: pool β†’ proj β†’ S^15 β†’ triangulate (768d)
  fine: per-pixel β†’ proj β†’ S^15 β†’ triangulate β†’ aggregate (768d)
  address = concat(coarse, fine) = 1536d

Stage 2 β€” CONDITION
  fuse(address + time_emb + class_emb + noise_emb) β†’ 1024d

Stage 3 β€” GENERATE
  4Γ— ResBlock(1024d) β†’ proj(16384d) β†’ reshape(256, 8, 8) β†’ decoder
```

## Files

### HuggingFace Integration
- `configuration_flow_match.py` β€” PretrainedConfig
- `modeling_flow_match.py` β€” PreTrainedModel (AutoModel compatible)

### Checkpoints (if present)
- `checkpoints/` β€” best checkpoints from each training run

### Samples (if present)
- `samples/` β€” v1 regulator samples
- `samples_bn/` β€” v2/v3 bottleneck samples
- `samples_cd/` β€” v3 pure constellation samples
- `samples_glfm/` β€” v4 GLFM samples

### Analysis Outputs (if present)
- `analysis/` β€” v1 analysis images
- `analysis_bn/` β€” v2 analysis images
- `analysis_cd/` β€” v3 analysis images
- `analysis_glfm/` β€” v4 analysis images

## Part of the GeoLIP Ecosystem

- [geolip-constellation-core](https://huggingface.co/AbstractPhil/geolip-constellation-core)
- [geolip-diffusion-proto](https://huggingface.co/AbstractPhil/geolip-diffusion-proto) (v1/v2 regulator)
- [geolip package](https://pypi.org/project/geolip/)
- [glip-autoencoder](https://github.com/AbstractEyes/glip-autoencoder)