geo-beatrix-fractal

Geometric Basin Classification for CIFAR-100

Immediate Assessment

The geo-beatrix variation is more capable at classification and inferior to the robust geometric capacity than the vit-beatrix transformer structure provides.

geo-beatrix has a different form of math and a new basin format entirely dependent on teaching traditional structures new behavior.

The system is hit or miss, and will be refined over time as the model family evolves.

The reality sets in when the classification gets higher, that this is a more capable model than a vit - and yet that vit I built has a far more robust set of tooling and capacity for learning transfer.

I'd say this is too big for standard classification tasks, and yet the classifier system does work somewhat - just not as well as SOTA.

Conclusion based on experimentation

This requires more experimentation on the subsystem before it can be utilized correctly. Optimizations need to happen to components, certain pieces need to be baselined to torch components for faster iterations. Even with loops removed this still has some issues with cantor stairs, but the batched stairs will be available on my repo today as well as the full model structure for the family of three here.

Alphamix and Fractalmix are hit-or-miss even with Cantor stairs, sometimes improving fidelity, sometimes reducing it.

Lacking attention mechanisms I consider this a resounding success as an experiment, and yet it fell short of resnet18 and resnet34 standalones - meaning the head only converted the math into something else, and fell short of the crossentropy goal.

That's okay though, I will refine the processes, improve the system, and return with additional trains for this version to further improve classifcation beyond the 69% chance - which may be HIGHER than the vit-beatrix, but it's considerably more shallow in comparison to geometric cohesion than the dual-stream transformer variation vit-beatrix-dualstream.

🚧 Training Concluded 🚧

Current Status: Idle


Current Performance

Metric Value
Best Test Accuracy 69.08%
Best Epoch 190
Current Ξ± (Cantor param) 0.4165
Total Parameters 45,161,489
Mixing Mode Fractal (triadic)

Architecture

  • Base: ResNet-style with residual blocks
  • Channels: 64 β†’ 128 β†’ 256 β†’ 512 β†’ 1024
  • Positional Encoding: Devil's Staircase (Cantor function, 1883)
  • PE Levels: 20
  • PE Features/Level: 4
  • Classification: Geometric Basin Compatibility
  • Attention: NONE
  • Cross-Entropy: NONE

Innovation

βœ… NO attention mechanisms
βœ… NO cross-entropy loss
βœ… Fractal positional encoding (Cantor function from 1883)
βœ… Geometric compatibility classification
βœ… Triadic fractal mixing (base-3 aligned)


Repository: https://huggingface.co/AbstractPhil/geo-beatrix
Author: AbstractPhil
Framework: PyTorch

Downloads last month

-

Downloads are not tracked for this model. How to track
Safetensors
Model size
60.5M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train AbstractPhil/geo-beatrix

Evaluation results