File size: 4,859 Bytes
ff647c3
 
 
a0d1b1d
eaccedc
07f2c45
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
 
 
 
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
 
b2bbba1
a0d1b1d
 
b2bbba1
a0d1b1d
 
eaccedc
a0d1b1d
 
 
 
 
 
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
 
 
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
 
 
eaccedc
 
 
 
 
 
 
a0d1b1d
 
eaccedc
a0d1b1d
eaccedc
 
 
a0d1b1d
 
 
 
 
eaccedc
 
a0d1b1d
 
eaccedc
a0d1b1d
 
 
 
eaccedc
a0d1b1d
eaccedc
a0d1b1d
eaccedc
 
a0d1b1d
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
eaccedc
 
ff647c3
eaccedc
a0d1b1d
 
 
 
 
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
 
 
 
eaccedc
a0d1b1d
 
 
 
eaccedc
a0d1b1d
 
 
 
 
eaccedc
a0d1b1d
eaccedc
a0d1b1d
eaccedc
a0d1b1d
 
 
 
eaccedc
 
 
 
a0d1b1d
 
 
 
 
 
eaccedc
 
 
 
 
a0d1b1d
eaccedc
 
 
a0d1b1d
 
 
 
 
eaccedc
ff647c3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
---
license: mit
---
## Model Description

**David Collective** is a geometric-simplex deep learning system that distills Stable Diffusion 1.5's knowledge into an ultra-efficient pentachoron-based architecture. This model was continued from epoch 20 to epoch 105, achieving remarkable performance with full pattern supervision.

### Architecture Highlights

- **Geometric Foundation**: Uses 5D pentachora (5-vertex simplices) instead of traditional attention
- **Multi-Scale Learning**: Extracts features from all 9 SD1.5 UNet blocks
- **Crystal Navigation**: 1000-class supervision (100 timesteps × 10 geometric patterns)
- **Parameter Efficiency**: Ultra-compact architecture with shared geometric structures
- **Full Supervision**: Every sample supervised by both timestep and geometric pattern

### Training Details

**Continuation Training:**
- Starting epoch: 20
- Final epoch: 105
- Total prompts trained: 600,500~ samples, 120,500~ prompts
- **All prompts included**: `prompts_all_epochs.jsonl` contains every prompt with metadata
- Dataset: Symbolic caption synthesis (complexity 1-5)
- Batch size: 128
- Learning rate: 1e-4 with cosine annealing
- Optimizer: AdamW (weight_decay=0.01)

**Final Metrics (Epoch 105):**
- Total Loss: 0.2923
- Timestep Accuracy: 66.98%
- Pattern Accuracy: 100.00%
- Full Accuracy: 66.98%
- Pattern Diversity: -0.221

### Active Blocks

David learns from all 9 SD1.5 UNet blocks:
- `down_0`, `down_1`, `down_2`, `down_3`: Coarse semantic features
- `mid`: Bottleneck representations
- `up_0`, `up_1`, `up_2`, `up_3`: Fine reconstruction details

### Loss Components

1. **Feature Similarity** (0.5): Cosine similarity with teacher
2. **Rose Loss** (0.3): Geometric alignment with crystal centroids
3. **Cross-Entropy** (0.2): 1000-class classification
4. **Pattern Diversity** (0.05): Encourages balanced pattern usage

## Usage

### Loading the Model

```python
import torch
from geovocab2.train.model.core.david_diffusion import DavidCollective, DavidCollectiveConfig
from safetensors.torch import load_file

# Load configuration
config = DavidCollectiveConfig(
    num_timestep_bins=100,
    num_feature_patterns_per_timestep=10,
    active_blocks=['down_0', 'down_1', 'down_2', 'down_3', 'mid', 'up_0', 'up_1', 'up_2', 'up_3'],
    david_sharing_mode='fully_shared',
    david_fusion_mode='deep_efficiency',
    use_belly=True,
    belly_expand=1.5
)

# Create model
model = DavidCollective(config)

# Load weights from safetensors
state_dict = load_file("model.safetensors")
model.load_state_dict(state_dict)
model.eval()

# Inference
with torch.no_grad():
    outputs = model(teacher_features, timesteps)
```

### Training Data

This model includes `prompts_all_epochs.jsonl` - every single prompt used during training with full metadata:

```json
{"timestamp": "2025-10-27T01:30:00", "epoch": 21, "batch": 0, "global_step": 6250, "sample_idx": 0, "timestep": 453, "timestep_bin": 45, "prompt": "a woman wearing red dress, against mountain landscape"}
```

**Total prompts:** 120,500 approximately

You can use this to:
- Analyze training data distribution
- Reproduce training
- Study prompt complexity vs model performance
- Generate similar synthetic datasets

## Technical Details

### Crystal System
- **Architecture**: Pentachoron-based geometric deep learning
- **Centroids**: 100 timestep bins × 10 patterns = 1000 anchors
- **Navigation**: Samples assigned to nearest pattern within timestep bin
- **Diversity**: Regularization prevents mode collapse

### Progressive Training
- Started with early blocks (down_0, down_1)
- Progressively activated all 9 blocks
- Each block warmed up for 2 epochs

### Pattern Supervision
Unlike traditional timestep-only supervision, David learns:
1. **When** (timestep bin 0-99)
2. **How** (geometric pattern 0-9 within that bin)
3. **Combined** (full 1000-class space)

This provides 10x finer-grained supervision of the diffusion process.

## Training History

Trained continuously from epoch 20 to epoch 105. See metrics:
- Timestep accuracy improved from ~60.3% to 66.98%
- Pattern accuracy maintained at 100.00%
- Loss decreased from 0.3431 to 0.2923

## Citation

```bibtex
@misc{david-collective-sd15,
  title={David Collective: Geometric Deep Learning for Diffusion Distillation},
  author={AbstractPhil},
  year={2025},
  publisher={HuggingFace},
  howpublished={\url{https://huggingface.co/AbstractPhil/david-collective-sd15-geometric-distillation}}
}
```

## License

MIT License - See repository for details.

## Acknowledgments

Built on the geometric deep learning research by AbstractPhil, using:
- Stable Diffusion 1.5 (teacher model)
- Pentachoron-based geometric algebra
- Crystalline consciousness architectures
- Symbolic caption synthesis

For more information, visit the [geovocab2 repository](https://github.com/AbstractEyes/lattice_vocabulary).