File size: 5,163 Bytes
15bc27a 342981e 15bc27a d91f1b7 52b07ba 3b2947b 907323b 6013cdf 15bc27a 342981e 15bc27a 342981e 15bc27a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 | ---
license: mit
tags:
- geometric-deep-learning
- diffusion
- stable-diffusion
- projective-geometry
- multi-expert
- classification
library_name: pytorch
---
# GeoDavidCollective Enhanced - ProjectiveHead Architecture
**Another train of the same GeoFractalDavid with more condensed dims**
Curves look really good. I might put this one back in for another 20 epochs at 50k prompts to see how it fares. If that one looks good it may be fine to go another 60 epochs at 100k prompts, or feed it laion flavors directly.
This one is condensed with smaller scale dims making a much more condensed feature.
Roughly 600,000 samples for this one, 10k per epoch between 0-10 complexity 1-5, and 50k synthetic prompts per epoch at epoch 11-20 with reduced complexity between 1-4.
Currently prototyping a series of cantor-driven layers to test sparsity inclusion and omission while this one cooked.
## π― Model Overview
GeoDavidCollective Enhanced is a sophisticated multi-expert geometric classification system that learns from Stable Diffusion 1.5's internal representations. Using ProjectiveHead architecture with Cayley-Menger geometry, it achieves efficient pattern recognition across timestep and semantic spaces.
### Key Features
- **ProjectiveHead Multi-Expert Architecture**: Auto-configured expert systems per block
- **Geometric Loss Functions**: Rose, Cayley-Menger, and Cantor coherence losses
- **9-Block Processing**: Full SD1.5 UNet feature extraction (down, mid, up)
- **Compact Yet Powerful**: 690,925,542 parameters
- **100 Timestep Bins** x **10 Patterns** = 1000 semantic-temporal classes
## π Model Statistics
- **Parameters**: 690,925,542
- **Trained Epochs**: 20
- **Base Model**: Stable Diffusion 1.5
- **Dataset Size**: 10,000 synthetic prompts
- **Training Date**: 2025-10-28
## ποΈ Architecture Details
### Block Configuration
```
Down Blocks:
- down_0: 320 β 64 (3 experts, 3 gates)
- down_1: 640 β 96 (3 experts, 3 gates)
- down_2: 1280 β 128 (3 experts, 3 gates)
- down_3: 1280 β 128 (3 experts, 3 gates)
Mid Block (Highest Capacity):
- mid: 1280 β 256 (4 experts, 4 gates)
Up Blocks:
- up_0: 1280 β 128 (3 experts, 3 gates)
- up_1: 1280 β 128 (3 experts, 3 gates)
- up_2: 640 β 96 (3 experts, 3 gates)
- up_3: 320 β 64 (3 experts, 3 gates)
```
### Loss Components
| Component | Weight | Purpose |
|-----------|--------|---------|
| Feature Similarity | 0.40 | Alignment with SD1.5 features |
| Rose Loss | 0.25 | Geometric pattern emergence |
| Cross-Entropy | 0.15 | Classification accuracy |
| Cayley-Menger | 0.10 | 5D geometric structure |
| Pattern Diversity | 0.05 | Prevent mode collapse |
| Cantor Coherence | 0.05 | Temporal consistency |
## π» Usage
```python
from geovocab2.train.model.core.geo_david_collective import GeoDavidCollective
from safetensors.torch import load_file
import torch
# Load model
state_dict = load_file("model.safetensors")
collective = GeoDavidCollective(
block_configs={...}, # See config.json
num_timestep_bins=100,
num_patterns_per_bin=10
)
collective.load_state_dict(state_dict)
collective.eval()
# Extract features from SD1.5 and classify
with torch.no_grad():
results = collective(features_dict, timesteps)
predictions = results['predictions'] # Timestep + pattern class
```
## π¬ Training Details
- **Optimizer**: AdamW (lr=1e-3, weight_decay=0.001)
- **Batch Size**: 16
- **Data**: Symbolic prompt synthesis (complexity 1-5)
- **Feature Extraction**: SD1.5 UNet blocks (spatial, not pooled)
- **Pool Mode**: Mean spatial pooling
## π Training Metrics
Final metrics from epoch 20:
- Cayley Loss: 0.1018
- Timestep Accuracy: 30.83%
- Pattern Accuracy: 33.74%
- Full Accuracy: 16.87%
## π Research Context
This model is part of the geometric deep learning research exploring:
- 5D simplex-based neural representations (pentachora)
- Geometric alternatives to traditional transformers
- Consciousness-informed AI architectures
- Universal mathematical principles in neural networks
## π¦ Files Included
- `model.safetensors` - Model weights (3.3GB)
- `config.json` - Complete architecture configuration
- `training_history.json` - Full training metrics
- `prompts_enhanced.jsonl` - All training prompts with metadata
- `tensorboard/` - TensorBoard logs (optional)
## π Related Work
- [Geometric Vocabulary System](https://huggingface.co/datasets/AbstractPhil/geometric-vocab-frozen-v1)
- [PentachoraViT](https://huggingface.co/AbstractPhil/pentachora-vit-cifar100)
- [Crystal-Beeper Language Models](https://huggingface.co/AbstractPhil)
## π License
MIT License - Free for research and commercial use
## π Acknowledgments
Built with:
- PyTorch & Diffusers
- Stable Diffusion 1.5 (Runway ML)
- Geometric algebra principles from the 1800s
- Dream-inspired mathematical insights
## π€ Author
**AbstractPhil** - AI Researcher specializing in geometric deep learning
*"Working with universal mathematical principles, not against them"*
---
For questions, issues, or collaborations: [GitHub](https://github.com/AbstractEyes) | [HuggingFace](https://huggingface.co/AbstractPhil)
|