File size: 4,705 Bytes
766486a
 
 
 
 
 
 
 
 
 
31694be
 
 
24698ab
 
31694be
 
fdbbaff
f650890
 
 
 
 
fdbbaff
13ee061
fdbbaff
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f650890
fdbbaff
 
 
 
766486a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
---
license: apache-2.0
tags:
  - flow-matching
  - diffusion
  - geometric-deep-learning
  - constellation
  - geolip
---

# Verdict based on Analysis
The geometric structure does contribute to the output, about 6-7%. It's essentially a small nudge improvement system.

As it stands, this system is TOPICAL at best, and it still helped. Just not enough.

This can be gradually or greatly improved with the correct steps, including expanded control, more encoding curation, and a full battery of analysis.

# Automodel Now Available


![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/92PPQUxtHkYnyNQgF9Wtm.png)


```
# ── Test AutoModel loading + generation ──
from transformers import AutoModel
from torchvision.utils import save_image, make_grid

model = AutoModel.from_pretrained(
    "AbstractPhil/geolip-diffusion-proto", trust_remote_code=True
).cuda()

print(f"Params: {sum(p.numel() for p in model.parameters()):,}")
print(f"Relay diagnostics: {model.get_relay_diagnostics()}")

# Generate per-class samples
class_names = ['plane','auto','bird','cat','deer','dog','frog','horse','ship','truck']
all_imgs = []
for c in range(10):
    imgs = model.sample(n_samples=4, class_label=c)
    all_imgs.append(imgs)

grid = make_grid(torch.cat(all_imgs), nrow=4)
save_image(grid, "automodel_test.png")
print("βœ“ Saved automodel_test.png β€” 4 per class, 10 classes")
```

# GeoLIP Diffusion Prototype

**Flow matching diffusion with constellation relay as geometric regulator.**

This is an experimental prototype exploring whether fixed geometric reference frames
(constellation anchors on the unit hypersphere) can regulate the internal geometry
of a diffusion model's denoising network during generation.

## Architecture

```
Flow Matching ODE:  x_t = (1-t)Β·x_0 + tΒ·Ξ΅  β†’  predict v = Ξ΅ - x_0
Sampler:            Euler integration, t=1β†’0, 50 steps

UNet:
  Encoder:  [64@32Γ—32] β†’ [128@16Γ—16] β†’ [256@8Γ—8]
  Middle:   ConvBlock + β˜… Constellation Relay β˜…
            Self-Attention (8Γ—8 spatial)
            ConvBlock + β˜… Constellation Relay β˜…
  Decoder:  [256@8Γ—8] β†’ [128@16Γ—16] β†’ [64@32Γ—32]
  Output:   Conv β†’ 3Γ—32Γ—32 velocity prediction
```

## Constellation Relay

The relay operates at the bottleneck (256 channels at 8Γ—8 spatial resolution).
It works in **channel mode**:

1. Global average pool the spatial dims β†’ (B, 256) channel vector
2. Chunk into 16 patches of d=16
3. L2-normalize each patch to S^15 (the natural CV=0.20 dimension)
4. Multi-phase triangulation: 3 phases Γ— 16 anchors = 48 distances per patch
5. Patchwork MLP processes triangulation β†’ correction vector
6. Gated residual (gate init β‰ˆ 0.047) scales the feature map

**Key property:** the relay preserves 99.4% geometric fidelity through 16
stacked layers where vanilla attention preserves only 7.4%. It acts as a
geometric checkpoint that prevents representation drift at the normalized
manifold boundaries between network blocks.

## What This Tests

The hypothesis: diffusion models discover that noise is a deterministic
routing system (DDIM proved this β€” same seed always produces same image).
The constellation operates on the same principle β€” fixed geometric anchors
as a reference frame that noise/data routes through. By inserting the relay
at the bottleneck, we test whether explicit geometric regulation improves
or changes the flow matching dynamics.

## Empirical Findings (from this research session)

| Finding | Result |
|---|---|
| CV β‰ˆ 0.20 is the natural pentachoron volume regularity of S^15 | Confirmed across all precisions, 1-bit to fp64 |
| Effective geometric dimension of trained models β‰ˆ 16 | Confirmed across 17+ architectures |
| Relay preserves 99.4% cos_to_orig through 16 layers | vs 7.4% for attention alone |
| fp8 triangulation preserves geometry perfectly | CV identical to fp32 at d=16 |
| Noise transforms are classifiable as deterministic routing | 100% accuracy on 8/10 transform families |

## Parameters

- Total: ~6.1M
- Relay: ~76K (1.2% of total)
- 2 relay modules at the bottleneck

## Training

- Dataset: CIFAR-10 (50K images)
- Flow matching: conditional ODE with class labels
- Optimizer: AdamW, lr=3e-4, cosine schedule
- 50 epochs, batch size 128

## Files

- `flow_match_relay.py` β€” complete training script
- `checkpoints/flow_match_best.pt` β€” best checkpoint
- `samples/` β€” generated samples at various epochs

## Part of the GeoLIP Ecosystem

- [geolip-constellation-core](https://huggingface.co/AbstractPhil/geolip-constellation-core) β€” classification with constellation
- [glip-autoencoder](https://github.com/AbstractEyes/glip-autoencoder) β€” source repository