File size: 4,705 Bytes
766486a 31694be 24698ab 31694be fdbbaff f650890 fdbbaff 13ee061 fdbbaff f650890 fdbbaff 766486a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 | ---
license: apache-2.0
tags:
- flow-matching
- diffusion
- geometric-deep-learning
- constellation
- geolip
---
# Verdict based on Analysis
The geometric structure does contribute to the output, about 6-7%. It's essentially a small nudge improvement system.
As it stands, this system is TOPICAL at best, and it still helped. Just not enough.
This can be gradually or greatly improved with the correct steps, including expanded control, more encoding curation, and a full battery of analysis.
# Automodel Now Available

```
# ββ Test AutoModel loading + generation ββ
from transformers import AutoModel
from torchvision.utils import save_image, make_grid
model = AutoModel.from_pretrained(
"AbstractPhil/geolip-diffusion-proto", trust_remote_code=True
).cuda()
print(f"Params: {sum(p.numel() for p in model.parameters()):,}")
print(f"Relay diagnostics: {model.get_relay_diagnostics()}")
# Generate per-class samples
class_names = ['plane','auto','bird','cat','deer','dog','frog','horse','ship','truck']
all_imgs = []
for c in range(10):
imgs = model.sample(n_samples=4, class_label=c)
all_imgs.append(imgs)
grid = make_grid(torch.cat(all_imgs), nrow=4)
save_image(grid, "automodel_test.png")
print("β Saved automodel_test.png β 4 per class, 10 classes")
```
# GeoLIP Diffusion Prototype
**Flow matching diffusion with constellation relay as geometric regulator.**
This is an experimental prototype exploring whether fixed geometric reference frames
(constellation anchors on the unit hypersphere) can regulate the internal geometry
of a diffusion model's denoising network during generation.
## Architecture
```
Flow Matching ODE: x_t = (1-t)Β·x_0 + tΒ·Ξ΅ β predict v = Ξ΅ - x_0
Sampler: Euler integration, t=1β0, 50 steps
UNet:
Encoder: [64@32Γ32] β [128@16Γ16] β [256@8Γ8]
Middle: ConvBlock + β
Constellation Relay β
Self-Attention (8Γ8 spatial)
ConvBlock + β
Constellation Relay β
Decoder: [256@8Γ8] β [128@16Γ16] β [64@32Γ32]
Output: Conv β 3Γ32Γ32 velocity prediction
```
## Constellation Relay
The relay operates at the bottleneck (256 channels at 8Γ8 spatial resolution).
It works in **channel mode**:
1. Global average pool the spatial dims β (B, 256) channel vector
2. Chunk into 16 patches of d=16
3. L2-normalize each patch to S^15 (the natural CV=0.20 dimension)
4. Multi-phase triangulation: 3 phases Γ 16 anchors = 48 distances per patch
5. Patchwork MLP processes triangulation β correction vector
6. Gated residual (gate init β 0.047) scales the feature map
**Key property:** the relay preserves 99.4% geometric fidelity through 16
stacked layers where vanilla attention preserves only 7.4%. It acts as a
geometric checkpoint that prevents representation drift at the normalized
manifold boundaries between network blocks.
## What This Tests
The hypothesis: diffusion models discover that noise is a deterministic
routing system (DDIM proved this β same seed always produces same image).
The constellation operates on the same principle β fixed geometric anchors
as a reference frame that noise/data routes through. By inserting the relay
at the bottleneck, we test whether explicit geometric regulation improves
or changes the flow matching dynamics.
## Empirical Findings (from this research session)
| Finding | Result |
|---|---|
| CV β 0.20 is the natural pentachoron volume regularity of S^15 | Confirmed across all precisions, 1-bit to fp64 |
| Effective geometric dimension of trained models β 16 | Confirmed across 17+ architectures |
| Relay preserves 99.4% cos_to_orig through 16 layers | vs 7.4% for attention alone |
| fp8 triangulation preserves geometry perfectly | CV identical to fp32 at d=16 |
| Noise transforms are classifiable as deterministic routing | 100% accuracy on 8/10 transform families |
## Parameters
- Total: ~6.1M
- Relay: ~76K (1.2% of total)
- 2 relay modules at the bottleneck
## Training
- Dataset: CIFAR-10 (50K images)
- Flow matching: conditional ODE with class labels
- Optimizer: AdamW, lr=3e-4, cosine schedule
- 50 epochs, batch size 128
## Files
- `flow_match_relay.py` β complete training script
- `checkpoints/flow_match_best.pt` β best checkpoint
- `samples/` β generated samples at various epochs
## Part of the GeoLIP Ecosystem
- [geolip-constellation-core](https://huggingface.co/AbstractPhil/geolip-constellation-core) β classification with constellation
- [glip-autoencoder](https://github.com/AbstractEyes/glip-autoencoder) β source repository
|