AbstractPhil commited on
Commit
61cee23
Β·
verified Β·
1 Parent(s): c452ff6

Epoch 50: 57.67%

Browse files
Files changed (1) hide show
  1. README.md +29 -30
README.md CHANGED
@@ -5,7 +5,7 @@ tags:
5
  - cifar100
6
  - geometric-learning
7
  - fractal-encoding
8
- - trained
9
  - no-attention
10
  - no-cross-entropy
11
  datasets:
@@ -34,9 +34,9 @@ model-index:
34
 
35
  **Geometric Basin Classification for CIFAR-100**
36
 
37
- πŸŽ‰ **Training Complete** πŸŽ‰
38
 
39
- Final Status: Epoch 200/200
40
 
41
  ---
42
 
@@ -46,23 +46,23 @@ Final Status: Epoch 200/200
46
  |--------|-------|
47
  | **Best Test Accuracy** | **66.72%** |
48
  | **Best Epoch** | 190 |
49
- | **Current Train Accuracy** | 45.28% |
50
- | **Current Test Accuracy** | 51.26% |
51
- | **Current Ξ± (Cantor param)** | 0.4260 |
52
- | **Total Parameters** | 45,172,297 |
53
- | **Training Time** | 0:50:42 |
54
 
55
  ### All Training Runs
56
 
57
  | Timestamp | Status | Best Epoch | Test Acc | Train Acc | Ξ± |
58
  |-----------|--------|------------|----------|-----------|---|
59
- | `20251010_004820` | βœ… | 190 | **66.72%** | 45.28% | 0.4260 |
60
 
61
  ### Comparison to State-of-the-Art
62
 
63
  | Model | Accuracy | Status |
64
  |-------|----------|--------|
65
- | **geo-beatrix (this model)** | **66.72%** | βœ… Complete |
66
  | vit-beatrix-dualstream | 66.0% | Vision Transformer + Cross-Entropy |
67
  | CLIP ViT-L/14 (zero-shot) | ~63-65% | 400M image-text pairs |
68
  | CLIP ViT-B/32 (zero-shot) | ~63.5% | Vision Transformer |
@@ -76,8 +76,8 @@ Final Status: Epoch 200/200
76
  - **Base**: ResNet-style with residual blocks
77
  - **Channels**: 64 β†’ 128 β†’ 256 β†’ 512 β†’ 1024
78
  - **Positional Encoding**: Devil's Staircase (Cantor function, 1883)
79
- - **PE Levels**: 5
80
- - **PE Features/Level**: 20
81
  - **Classification**: Geometric Basin Compatibility (NO cross-entropy)
82
  - **Attention Mechanisms**: NONE
83
 
@@ -95,8 +95,8 @@ Final Status: Epoch 200/200
95
  "base_learning_rate": 0.001,
96
  "weight_decay": 0.05,
97
  "warmup_epochs": 10,
98
- "pe_levels": 5,
99
- "pe_features_per_level": 20,
100
  "dropout": 0.1,
101
  "upload_every_n_epochs": 50,
102
  "alphamix": {
@@ -113,7 +113,7 @@ Final Status: Epoch 200/200
113
  "loss_function": "Geometric Basin Compatibility",
114
  "cross_entropy": false,
115
  "attention_mechanisms": false,
116
- "timestamp": "20251010_004820"
117
  }
118
  ```
119
 
@@ -127,7 +127,7 @@ Final Status: Epoch 200/200
127
  β”œβ”€β”€ best_model_info.json (which epoch/run this came from)
128
  β”œβ”€β”€ runs_history.json (all training runs and their results)
129
  β”œβ”€β”€ README.md
130
- β”œβ”€β”€ weights/geo-beatrix/20251010_004820/
131
  β”‚ β”œβ”€β”€ model.pt (best from this training run)
132
  β”‚ β”œβ”€β”€ model.safetensors (best from this training run)
133
  β”‚ β”œβ”€β”€ config.json
@@ -137,7 +137,7 @@ Final Status: Epoch 200/200
137
  β”‚ β”œβ”€β”€ checkpoint_epoch_20.safetensors
138
  β”‚ └── checkpoint_epoch_30.safetensors
139
  β”‚ (snapshots every 50 epochs)
140
- └── runs/geo-beatrix/20251010_004820/
141
  β”œβ”€β”€ events.out.tfevents.* (TensorBoard logs)
142
  └── metrics.csv (training metrics)
143
  ```
@@ -173,13 +173,13 @@ with open(info_path) as f:
173
  # Or download from specific training run
174
  model_path = hf_hub_download(
175
  repo_id="AbstractPhil/geo-beatrix",
176
- filename="weights/geo-beatrix/20251010_004820/model.safetensors"
177
  )
178
 
179
  # Download specific epoch checkpoint
180
  epoch_checkpoint = hf_hub_download(
181
  repo_id="AbstractPhil/geo-beatrix",
182
- filename="weights/geo-beatrix/20251010_004820/checkpoints/checkpoint_epoch_100.safetensors"
183
  )
184
  ```
185
 
@@ -189,22 +189,21 @@ epoch_checkpoint = hf_hub_download(
189
 
190
  ### Best Checkpoint
191
  - Epoch: 190
192
- - Train Acc: 51.34%
193
- - Test Acc: 56.68%
194
- - Alpha: 0.4259
195
- - Loss: 0.8149
196
 
197
  ### Latest 5 Epochs
198
 
199
- - **Epoch 196**: Train 49.67%, Test 0.00%, Ξ±=0.4260, Loss=0.8037
200
- - **Epoch 197**: Train 46.56%, Test 0.00%, Ξ±=0.4260, Loss=0.7564
201
- - **Epoch 198**: Train 50.45%, Test 0.00%, Ξ±=0.4260, Loss=0.8134
202
- - **Epoch 199**: Train 53.19%, Test 0.00%, Ξ±=0.4260, Loss=0.8330
203
- - **Epoch 200**: Train 45.28%, Test 51.26%, Ξ±=0.4260, Loss=0.7648
204
 
205
  ### Training Milestones
206
- - 🎯 **50% Accuracy** reached at epoch 30
207
- - 🎯 **60% Accuracy** reached at epoch 95
208
  - πŸ“Š **Ξ± β‰₯ 0.40** reached at epoch 9
209
 
210
  ---
 
5
  - cifar100
6
  - geometric-learning
7
  - fractal-encoding
8
+ - in-training
9
  - no-attention
10
  - no-cross-entropy
11
  datasets:
 
34
 
35
  **Geometric Basin Classification for CIFAR-100**
36
 
37
+ 🚧 **Training in Progress** 🚧
38
 
39
+ Current Status: Epoch 50/200
40
 
41
  ---
42
 
 
46
  |--------|-------|
47
  | **Best Test Accuracy** | **66.72%** |
48
  | **Best Epoch** | 190 |
49
+ | **Current Train Accuracy** | 61.38% |
50
+ | **Current Test Accuracy** | 57.67% |
51
+ | **Current Ξ± (Cantor param)** | 0.4204 |
52
+ | **Total Parameters** | 45,235,067 |
53
+ | **Training Time** | 0:13:10 |
54
 
55
  ### All Training Runs
56
 
57
  | Timestamp | Status | Best Epoch | Test Acc | Train Acc | Ξ± |
58
  |-----------|--------|------------|----------|-----------|---|
59
+ | `20251010_014649` | πŸ”„ | 190 | **66.72%** | 61.38% | 0.4204 |
60
 
61
  ### Comparison to State-of-the-Art
62
 
63
  | Model | Accuracy | Status |
64
  |-------|----------|--------|
65
+ | **geo-beatrix (this model)** | **66.72%** | πŸ”„ Training |
66
  | vit-beatrix-dualstream | 66.0% | Vision Transformer + Cross-Entropy |
67
  | CLIP ViT-L/14 (zero-shot) | ~63-65% | 400M image-text pairs |
68
  | CLIP ViT-B/32 (zero-shot) | ~63.5% | Vision Transformer |
 
76
  - **Base**: ResNet-style with residual blocks
77
  - **Channels**: 64 β†’ 128 β†’ 256 β†’ 512 β†’ 1024
78
  - **Positional Encoding**: Devil's Staircase (Cantor function, 1883)
79
+ - **PE Levels**: 20
80
+ - **PE Features/Level**: 10
81
  - **Classification**: Geometric Basin Compatibility (NO cross-entropy)
82
  - **Attention Mechanisms**: NONE
83
 
 
95
  "base_learning_rate": 0.001,
96
  "weight_decay": 0.05,
97
  "warmup_epochs": 10,
98
+ "pe_levels": 20,
99
+ "pe_features_per_level": 10,
100
  "dropout": 0.1,
101
  "upload_every_n_epochs": 50,
102
  "alphamix": {
 
113
  "loss_function": "Geometric Basin Compatibility",
114
  "cross_entropy": false,
115
  "attention_mechanisms": false,
116
+ "timestamp": "20251010_014649"
117
  }
118
  ```
119
 
 
127
  β”œβ”€β”€ best_model_info.json (which epoch/run this came from)
128
  β”œβ”€β”€ runs_history.json (all training runs and their results)
129
  β”œβ”€β”€ README.md
130
+ β”œβ”€β”€ weights/geo-beatrix/20251010_014649/
131
  β”‚ β”œβ”€β”€ model.pt (best from this training run)
132
  β”‚ β”œβ”€β”€ model.safetensors (best from this training run)
133
  β”‚ β”œβ”€β”€ config.json
 
137
  β”‚ β”œβ”€β”€ checkpoint_epoch_20.safetensors
138
  β”‚ └── checkpoint_epoch_30.safetensors
139
  β”‚ (snapshots every 50 epochs)
140
+ └── runs/geo-beatrix/20251010_014649/
141
  β”œβ”€β”€ events.out.tfevents.* (TensorBoard logs)
142
  └── metrics.csv (training metrics)
143
  ```
 
173
  # Or download from specific training run
174
  model_path = hf_hub_download(
175
  repo_id="AbstractPhil/geo-beatrix",
176
+ filename="weights/geo-beatrix/20251010_014649/model.safetensors"
177
  )
178
 
179
  # Download specific epoch checkpoint
180
  epoch_checkpoint = hf_hub_download(
181
  repo_id="AbstractPhil/geo-beatrix",
182
+ filename="weights/geo-beatrix/20251010_014649/checkpoints/checkpoint_epoch_100.safetensors"
183
  )
184
  ```
185
 
 
189
 
190
  ### Best Checkpoint
191
  - Epoch: 190
192
+ - Train Acc: 61.38%
193
+ - Test Acc: 66.72%
194
+ - Alpha: 0.4204
195
+ - Loss: 0.0000
196
 
197
  ### Latest 5 Epochs
198
 
199
+ - **Epoch 46**: Train 59.76%, Test 0.00%, Ξ±=0.4153, Loss=1.2259
200
+ - **Epoch 47**: Train 59.94%, Test 0.00%, Ξ±=0.4171, Loss=1.1919
201
+ - **Epoch 48**: Train 61.30%, Test 0.00%, Ξ±=0.4238, Loss=1.2102
202
+ - **Epoch 49**: Train 61.85%, Test 0.00%, Ξ±=0.4170, Loss=1.1685
203
+ - **Epoch 50**: Train 61.38%, Test 57.67%, Ξ±=0.4204, Loss=1.1308
204
 
205
  ### Training Milestones
206
+ - 🎯 **50% Accuracy** reached at epoch 40
 
207
  - πŸ“Š **Ξ± β‰₯ 0.40** reached at epoch 9
208
 
209
  ---