AbstractPhil's picture
geovocab v1: schnell_full_1_5e-5 — 1 epoch 50k synthetic-characters
58b21ed verified
---
license: mit
library_name: sd15-flow-trainer
tags:
- geometric-deep-learning
- stable-diffusion
- ksimplex
- pentachoron
- flow-matching
- cross-attention-prior
base_model: sd-legacy/stable-diffusion-v1-5
pipeline_tag: text-to-image
---
# KSimplex Geometric Attention Prior
Geometric cross-attention prior for SD1.5 using pentachoron (4-simplex) structures.
## Architecture
| Component | Params |
|-----------|--------|
| SD1.5 UNet (frozen) | 859,520,964 |
| **Geo prior (trained)** | **4,845,725** |
| **Geo conditioner (trained)** | **1,613,847** |
## Simplex Configuration
| Parameter | Value |
|-----------|-------|
| k (simplex dim) | 4 |
| Embedding dim | 32 |
| Feature dim | 768 |
| Stacked layers | 4 |
| Attention heads | 8 |
| Base deformation | 0.25 |
| Residual blend | learnable |
| Timestep conditioned | True |
## GeoVocab Conditioning
| Parameter | Value |
|-----------|-------|
| Gate dim | 17 |
| Patch feat dim | 256 |
| Num patches | 64 |
| Cross-attention | enabled |
| Cross-attn heads | 8 |
| Blend mode | learnable |
## Usage
```python
from sd15_trainer_geo.pipeline import load_pipeline
pipe = load_pipeline(geo_repo_id="AbstractPhil/sd15-geovocab-lora-prototype")
```
## Training Info
- **dataset**: AbstractPhil/synthetic-characters (schnell_full_1_512)
- **subdir**: schnell_full_1_5e-5
- **samples**: 50000
- **epochs**: 10
- **steps**: 83330
- **shift**: 2.0
- **base_lr**: 5e-05
- **min_snr_gamma**: 5.0
- **cfg_dropout**: 0.1
- **batch_size**: 6
- **geo_loss_weight**: 0.01
- **geovocab_lr_mult**: 2.0
- **clip_vae**: AbstractPhil/geovae-proto/clip_vae/best_model.pt
- **patch_maker**: AbstractPhil/geovocab-patch-maker
- **loss_final**: 0.3035515168607235
## License
MIT — [AbstractPhil](https://huggingface.co/AbstractPhil)