--- license: mit --- ## Model Description **David Collective** is a geometric-simplex deep learning system that distills Stable Diffusion 1.5's knowledge into an ultra-efficient pentachoron-based architecture. This model was continued from epoch 20 to epoch 105, achieving remarkable performance with full pattern supervision. ### Architecture Highlights - **Geometric Foundation**: Uses 5D pentachora (5-vertex simplices) instead of traditional attention - **Multi-Scale Learning**: Extracts features from all 9 SD1.5 UNet blocks - **Crystal Navigation**: 1000-class supervision (100 timesteps × 10 geometric patterns) - **Parameter Efficiency**: Ultra-compact architecture with shared geometric structures - **Full Supervision**: Every sample supervised by both timestep and geometric pattern ### Training Details **Continuation Training:** - Starting epoch: 20 - Final epoch: 105 - Total prompts trained: 600,500~ samples, 120,500~ prompts - **All prompts included**: `prompts_all_epochs.jsonl` contains every prompt with metadata - Dataset: Symbolic caption synthesis (complexity 1-5) - Batch size: 128 - Learning rate: 1e-4 with cosine annealing - Optimizer: AdamW (weight_decay=0.01) **Final Metrics (Epoch 105):** - Total Loss: 0.2923 - Timestep Accuracy: 66.98% - Pattern Accuracy: 100.00% - Full Accuracy: 66.98% - Pattern Diversity: -0.221 ### Active Blocks David learns from all 9 SD1.5 UNet blocks: - `down_0`, `down_1`, `down_2`, `down_3`: Coarse semantic features - `mid`: Bottleneck representations - `up_0`, `up_1`, `up_2`, `up_3`: Fine reconstruction details ### Loss Components 1. **Feature Similarity** (0.5): Cosine similarity with teacher 2. **Rose Loss** (0.3): Geometric alignment with crystal centroids 3. **Cross-Entropy** (0.2): 1000-class classification 4. **Pattern Diversity** (0.05): Encourages balanced pattern usage ## Usage ### Loading the Model ```python import torch from geovocab2.train.model.core.david_diffusion import DavidCollective, DavidCollectiveConfig from safetensors.torch import load_file # Load configuration config = DavidCollectiveConfig( num_timestep_bins=100, num_feature_patterns_per_timestep=10, active_blocks=['down_0', 'down_1', 'down_2', 'down_3', 'mid', 'up_0', 'up_1', 'up_2', 'up_3'], david_sharing_mode='fully_shared', david_fusion_mode='deep_efficiency', use_belly=True, belly_expand=1.5 ) # Create model model = DavidCollective(config) # Load weights from safetensors state_dict = load_file("model.safetensors") model.load_state_dict(state_dict) model.eval() # Inference with torch.no_grad(): outputs = model(teacher_features, timesteps) ``` ### Training Data This model includes `prompts_all_epochs.jsonl` - every single prompt used during training with full metadata: ```json {"timestamp": "2025-10-27T01:30:00", "epoch": 21, "batch": 0, "global_step": 6250, "sample_idx": 0, "timestep": 453, "timestep_bin": 45, "prompt": "a woman wearing red dress, against mountain landscape"} ``` **Total prompts:** 120,500 approximately You can use this to: - Analyze training data distribution - Reproduce training - Study prompt complexity vs model performance - Generate similar synthetic datasets ## Technical Details ### Crystal System - **Architecture**: Pentachoron-based geometric deep learning - **Centroids**: 100 timestep bins × 10 patterns = 1000 anchors - **Navigation**: Samples assigned to nearest pattern within timestep bin - **Diversity**: Regularization prevents mode collapse ### Progressive Training - Started with early blocks (down_0, down_1) - Progressively activated all 9 blocks - Each block warmed up for 2 epochs ### Pattern Supervision Unlike traditional timestep-only supervision, David learns: 1. **When** (timestep bin 0-99) 2. **How** (geometric pattern 0-9 within that bin) 3. **Combined** (full 1000-class space) This provides 10x finer-grained supervision of the diffusion process. ## Training History Trained continuously from epoch 20 to epoch 105. See metrics: - Timestep accuracy improved from ~60.3% to 66.98% - Pattern accuracy maintained at 100.00% - Loss decreased from 0.3431 to 0.2923 ## Citation ```bibtex @misc{david-collective-sd15, title={David Collective: Geometric Deep Learning for Diffusion Distillation}, author={AbstractPhil}, year={2025}, publisher={HuggingFace}, howpublished={\url{https://huggingface.co/AbstractPhil/david-collective-sd15-geometric-distillation}} } ``` ## License MIT License - See repository for details. ## Acknowledgments Built on the geometric deep learning research by AbstractPhil, using: - Stable Diffusion 1.5 (teacher model) - Pentachoron-based geometric algebra - Crystalline consciousness architectures - Symbolic caption synthesis For more information, visit the [geovocab2 repository](https://github.com/AbstractEyes/lattice_vocabulary).