sd15-geoflow-test-44-1_000 β€” Geometric Prior Burn Test

Extreme-repetition burn test of the KSimplex geometric prior on 22 source images for 10,000 total steps. This experiment tests whether the pentachoron-based geometric prior can learn subject-specific anchoring from minimal data with high repetition β€” analogous to LoRA training but operating in simplex coordinate space rather than weight delta space.

What This Is

A 4.8M parameter geometric prior (0.56% of the frozen 860M SD1.5 UNet) trained on 22 images of a single subject repeated 454Γ— each across 6 aspect-ratio buckets, totaling ~10k samples over ~10k training steps.

The source images are promotional and fan-sourced photographs of actress Terry Farrell as Jadzia Dax from Star Trek: Deep Space Nine.

This answers crucial questions

This is testing the potential for an actual 1000 step lora with small amounts of images. The results are mixed so far.

Architecture

  • Base: SD1.5 UNet (frozen) with Lune rectified flow weights from AbstractPhil/tinyflux-experts
  • Trainable: KSimplex cross-attention prior β€” 4 layers of pentachoron-based geometric attention
  • Flow: Rectified flow matching with logit-normal timestep sampling, shift=2.5
  • Bucketing: 6 AR buckets (576Γ—448, 384Γ—640, 448Γ—576, 704Γ—384, 384Γ—704, 512Γ—512)

Training Details

Bucket Images Samples Steps Final Loss
576Γ—448 11 4,994 5,000 0.312
384Γ—640 4 1,816 1,818 0.284
448Γ—576 3 1,362 1,363 0.338
704Γ—384 2 908 909 0.216
384Γ—704 1 454 500 0.287
512Γ—512 1 454 500 0.321

Training ran sequentially per bucket with shared geo_prior weights carrying over. Batch size 6, base LR 5e-5 with cosine decay.

Post-Training Analysis

Weight Inspection

Metric Value Interpretation
Blend Ξ² 0.4991 Nearly 50/50 CLIP vs geometric β€” barely moved from init
L0 Ξ΄ 0.268 Minimal deformation
L1 Ξ΄ 0.272 Minimal deformation
L2 Ξ΄ 0.296 Slight increase
L3 Ξ΄ 0.260 Below init (0.25 base + learned offset)

Vertex Weight Entropy (the key signal)

Layer Trained Fresh Ξ”
L0 1.536 1.511 +0.026
L1 1.330 1.537 βˆ’0.208
L2 1.423 1.526 βˆ’0.103
L3 1.553 1.560 βˆ’0.008

Comparison across all trained models:

Model L3 vw_entropy L3 Ξ” from fresh Character
Object-relations (50k) 0.243 βˆ’1.291 Hard vertex routing
ImageNet (10k) 0.749 βˆ’0.868 Moderate sharpening
Characters (50k) 1.277 βˆ’0.265 Soft attribute binding
Burn test (22 imgs - 1000 steps) 1.553 βˆ’0.008 Slightly established

The burn test shows the geometric prior barely moved from initialization despite 10k steps and 454Γ— repetition. This is the expected result for 22 images of a single subject β€” there isn't enough compositional diversity to drive simplex specialization.

Branch Point

Recommended branch point: t β‰ˆ 0.68 (much later than object-relations at 0.31 or characters at 0.28).

Analysis Artifacts

Training Analysis

Post-Training Analysis

Sample Images

Key Findings

  1. 22 images is below the diversity threshold for geometric specialization. The prior needs compositional variety to develop distinct vertex routing strategies. A single repeated subject gives it nothing to differentiate geometrically.

  2. The geo loss converged to βˆ’0.047, well below the βˆ’0.053 seen in object-relations and characters. The simplex found a valid geometric configuration but a less structured one.

  3. L1 showed the most movement (vw_entropy Ξ” = βˆ’0.208), suggesting the prior concentrated subject identity information at the second layer while leaving other layers near-uniform.

  4. Blend stayed at 0.499 β€” the prior learned almost nothing worth asserting over CLIP. In contrast, object-relations moved blend to 0.472 and characters to 0.476.

  5. This is NOT a failure β€” it demonstrates that the geometric prior correctly identifies when there isn't enough structural signal to justify specialization, unlike LoRA which would overfit the weight deltas regardless.

Usage

from sd15_trainer_geo.pipeline import load_pipeline

pipe = load_pipeline(
    geo_repo_id="AbstractPhil/sd15-geoflow-test-44",
    device="cuda",
    dtype=torch.float16,
)
pipe.unet.load_pretrained(
    "AbstractPhil/tinyflux-experts",
    subfolder="",
    filename="sd15-flow-lune-unet.safetensors",
)

License

Dual license:

  • Code, architecture, and training methodology: MIT License. The KSimplex geometric prior architecture, training scripts, and analysis tools are freely available for any use.

  • Weights and derivative images: The trained weights encode information derived from copyrighted promotional imagery from Star Trek: Deep Space Nine (Paramount/CBS). Generated images from these weights may reproduce likenesses subject to copyright and publicity rights. Users are responsible for ensuring their use of generated imagery complies with applicable intellectual property law. These weights are provided for research and non-commercial use only.

Part of the Geometric Prior Research Series

Model Dataset Samples Steps L3 vw_entropy
sd15-rectified-geometric-matching ImageNet subset 10,000 10,000 0.749
sd15-geoflow-object-association Object relations 50,000 8,333 0.243
sd15-geoflow-characters Synthetic characters 50,000 8,333 1.277
sd15-geoflow-test-44 Single subject (burn) 9,988 10,090 1.553
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for AbstractPhil/sd15-geoflow-test-44-1000

Finetuned
(345)
this model