geo-beatrix / README.md
AbstractPhil's picture
Update README.md
c037576 verified
metadata
license: mit
tags:
  - image-classification
  - cifar100
  - geometric-learning
  - fractal-encoding
  - in-training
  - no-attention
  - no-cross-entropy
datasets:
  - cifar100
metrics:
  - accuracy
library_name: pytorch
pipeline_tag: image-classification
model-index:
  - name: geo-beatrix-fractal
    results:
      - task:
          type: image-classification
          name: Image Classification
        dataset:
          name: CIFAR-100
          type: cifar100
        metrics:
          - type: accuracy
            value: 69.08
            name: Test Accuracy
            verified: false

geo-beatrix-fractal

Geometric Basin Classification for CIFAR-100

Immediate Assessment

The geo-beatrix variation is more capable at classification and inferior to the robust geometric capacity than the vit-beatrix transformer structure provides.

geo-beatrix has a different form of math and a new basin format entirely dependent on teaching traditional structures new behavior.

The system is hit or miss, and will be refined over time as the model family evolves.

The reality sets in when the classification gets higher, that this is a more capable model than a vit - and yet that vit I built has a far more robust set of tooling and capacity for learning transfer.

I'd say this is too big for standard classification tasks, and yet the classifier system does work somewhat - just not as well as SOTA.

Conclusion based on experimentation

This requires more experimentation on the subsystem before it can be utilized correctly. Optimizations need to happen to components, certain pieces need to be baselined to torch components for faster iterations. Even with loops removed this still has some issues with cantor stairs, but the batched stairs will be available on my repo today as well as the full model structure for the family of three here.

Alphamix and Fractalmix are hit-or-miss even with Cantor stairs, sometimes improving fidelity, sometimes reducing it.

Lacking attention mechanisms I consider this a resounding success as an experiment, and yet it fell short of resnet18 and resnet34 standalones - meaning the head only converted the math into something else, and fell short of the crossentropy goal.

That's okay though, I will refine the processes, improve the system, and return with additional trains for this version to further improve classifcation beyond the 69% chance - which may be HIGHER than the vit-beatrix, but it's considerably more shallow in comparison to geometric cohesion than the dual-stream transformer variation vit-beatrix-dualstream.

🚧 Training Concluded 🚧

Current Status: Idle


Current Performance

Metric Value
Best Test Accuracy 69.08%
Best Epoch 190
Current Ξ± (Cantor param) 0.4165
Total Parameters 45,161,489
Mixing Mode Fractal (triadic)

Architecture

  • Base: ResNet-style with residual blocks
  • Channels: 64 β†’ 128 β†’ 256 β†’ 512 β†’ 1024
  • Positional Encoding: Devil's Staircase (Cantor function, 1883)
  • PE Levels: 20
  • PE Features/Level: 4
  • Classification: Geometric Basin Compatibility
  • Attention: NONE
  • Cross-Entropy: NONE

Innovation

βœ… NO attention mechanisms
βœ… NO cross-entropy loss
βœ… Fractal positional encoding (Cantor function from 1883)
βœ… Geometric compatibility classification
βœ… Triadic fractal mixing (base-3 aligned)


Repository: https://huggingface.co/AbstractPhil/geo-beatrix
Author: AbstractPhil
Framework: PyTorch