AbstractPhil commited on
Commit
6e4c74c
Β·
verified Β·
1 Parent(s): 93e2ae0

Epoch 50: 58.38%

Browse files
Files changed (1) hide show
  1. README.md +62 -61
README.md CHANGED
@@ -5,7 +5,7 @@ tags:
5
  - cifar100
6
  - geometric-learning
7
  - fractal-encoding
8
- - trained
9
  - no-attention
10
  - no-cross-entropy
11
  datasets:
@@ -25,7 +25,7 @@ model-index:
25
  type: cifar100
26
  metrics:
27
  - type: accuracy
28
- value: 67.00
29
  name: Test Accuracy
30
  verified: false
31
  ---
@@ -34,9 +34,9 @@ model-index:
34
 
35
  **Geometric Basin Classification for CIFAR-100**
36
 
37
- πŸŽ‰ **Training Complete** πŸŽ‰
38
 
39
- Final Status: Epoch 200/200
40
 
41
  ---
42
 
@@ -44,24 +44,24 @@ Final Status: Epoch 200/200
44
 
45
  | Metric | Value |
46
  |--------|-------|
47
- | **Best Test Accuracy** | **67.00%** |
48
- | **Best Epoch** | 190 |
49
- | **Current Train Accuracy** | 63.49% |
50
- | **Current Test Accuracy** | 64.32% |
51
- | **Current Ξ± (Cantor param)** | 0.4422 |
52
- | **Total Parameters** | 45,161,489 |
53
- | **Training Time** | 1:10:27 |
54
 
55
  ### Comparison to State-of-the-Art
56
 
57
  | Model | Accuracy | Status |
58
  |-------|----------|--------|
59
- | **geo-beatrix (this model)** | **67.00%** | βœ… Complete |
60
  | vit-beatrix-dualstream | 66.0% | Vision Transformer + Cross-Entropy |
61
  | CLIP ViT-L/14 (zero-shot) | ~63-65% | 400M image-text pairs |
62
  | CLIP ViT-B/32 (zero-shot) | ~63.5% | Vision Transformer |
63
 
64
- βœ… **geo-beatrix has surpassed all baselines!**
65
 
66
  ---
67
 
@@ -71,7 +71,7 @@ Final Status: Epoch 200/200
71
  - **Channels**: 64 β†’ 128 β†’ 256 β†’ 512 β†’ 1024
72
  - **Positional Encoding**: Devil's Staircase (Cantor function, 1883)
73
  - **PE Levels**: 20
74
- - **PE Features/Level**: 4
75
  - **Classification**: Geometric Basin Compatibility (NO cross-entropy)
76
  - **Attention Mechanisms**: NONE
77
 
@@ -90,9 +90,9 @@ Final Status: Epoch 200/200
90
  "weight_decay": 0.05,
91
  "warmup_epochs": 10,
92
  "pe_levels": 20,
93
- "pe_features_per_level": 4,
94
  "dropout": 0.1,
95
- "upload_every_n_epochs": 10,
96
  "alphamix": {
97
  "enabled": true,
98
  "range": [
@@ -107,7 +107,7 @@ Final Status: Epoch 200/200
107
  "loss_function": "Geometric Basin Compatibility",
108
  "cross_entropy": false,
109
  "attention_mechanisms": false,
110
- "timestamp": "20251009_222052"
111
  }
112
  ```
113
 
@@ -116,22 +116,27 @@ Final Status: Epoch 200/200
116
  ## Files Structure
117
 
118
  ```
119
- weights/geo-beatrix/20251009_222052/
120
- β”œβ”€β”€ model.pt (best checkpoint - PyTorch)
121
- β”œβ”€β”€ model.safetensors (best checkpoint - SafeTensors)
122
- β”œβ”€β”€ config.json (model configuration)
123
- β”œβ”€β”€ training_log.txt (training log)
124
- └── checkpoints/
125
- β”œβ”€β”€ checkpoint_epoch_10.safetensors
126
- β”œβ”€β”€ checkpoint_epoch_20.safetensors
127
- └── checkpoint_epoch_30.safetensors
128
- (snapshots every 10 epochs)
129
-
130
- runs/geo-beatrix/20251009_222052/
131
- β”œβ”€β”€ events.out.tfevents.* (TensorBoard logs)
132
- └── metrics.csv (training metrics)
 
 
 
133
  ```
134
 
 
 
135
  ---
136
 
137
  ## Usage
@@ -140,34 +145,34 @@ runs/geo-beatrix/20251009_222052/
140
  from huggingface_hub import hf_hub_download
141
  import torch
142
 
143
- # Download best model (SafeTensors - recommended)
144
  from safetensors.torch import load_file
145
  model_path = hf_hub_download(
146
  repo_id="AbstractPhil/geo-beatrix",
147
- filename="weights/geo-beatrix/20251009_222052/model.safetensors"
148
  )
149
  state_dict = load_file(model_path)
150
  # model.load_state_dict(state_dict)
151
 
152
- # Or download PyTorch checkpoint (includes optimizer state)
153
- checkpoint_path = hf_hub_download(
154
  repo_id="AbstractPhil/geo-beatrix",
155
- filename="weights/geo-beatrix/20251009_222052/model.pt"
156
  )
157
- checkpoint = torch.load(checkpoint_path)
158
- # model.load_state_dict(checkpoint['model_state_dict'])
 
159
 
160
- # Download specific epoch checkpoint
161
- epoch_checkpoint = hf_hub_download(
162
  repo_id="AbstractPhil/geo-beatrix",
163
- filename="weights/geo-beatrix/20251009_222052/checkpoints/checkpoint_epoch_100.safetensors"
164
  )
165
 
166
- # Download TensorBoard logs for visualization
167
- import glob
168
- tensorboard_files = hf_hub_download(
169
  repo_id="AbstractPhil/geo-beatrix",
170
- filename="runs/geo-beatrix/20251009_222052/events.out.tfevents.*"
171
  )
172
  ```
173
 
@@ -176,27 +181,23 @@ tensorboard_files = hf_hub_download(
176
  ## Training History
177
 
178
  ### Best Checkpoint
179
- - Epoch: 190
180
- - Train Acc: 61.47%
181
- - Test Acc: 67.00%
182
- - Alpha: 0.4425
183
- - Loss: 0.5031
184
 
185
  ### Latest 5 Epochs
186
 
187
- - **Epoch 196**: Train 63.32%, Test 0.00%, Ξ±=0.4422, Loss=0.4936
188
- - **Epoch 197**: Train 65.20%, Test 0.00%, Ξ±=0.4422, Loss=0.5069
189
- - **Epoch 198**: Train 60.80%, Test 0.00%, Ξ±=0.4422, Loss=0.4929
190
- - **Epoch 199**: Train 58.87%, Test 0.00%, Ξ±=0.4422, Loss=0.4874
191
- - **Epoch 200**: Train 63.49%, Test 64.32%, Ξ±=0.4422, Loss=0.5049
192
 
193
  ### Training Milestones
194
- - 🎯 **50% Accuracy** reached at epoch 45
195
- - 🎯 **60% Accuracy** reached at epoch 70
196
- - πŸ† **Beat vit-beatrix (66.0%)** at epoch 140
197
- - πŸš€ **67% Accuracy** reached at epoch 190
198
- - πŸ“Š **Ξ± β‰₯ 0.40** reached at epoch 10
199
- - πŸ“Š **Ξ± β‰₯ 0.44** (near triadic equilibrium) at epoch 116
200
 
201
  ---
202
 
 
5
  - cifar100
6
  - geometric-learning
7
  - fractal-encoding
8
+ - in-training
9
  - no-attention
10
  - no-cross-entropy
11
  datasets:
 
25
  type: cifar100
26
  metrics:
27
  - type: accuracy
28
+ value: 58.38
29
  name: Test Accuracy
30
  verified: false
31
  ---
 
34
 
35
  **Geometric Basin Classification for CIFAR-100**
36
 
37
+ 🚧 **Training in Progress** 🚧
38
 
39
+ Current Status: Epoch 50/200
40
 
41
  ---
42
 
 
44
 
45
  | Metric | Value |
46
  |--------|-------|
47
+ | **Best Test Accuracy** | **58.38%** |
48
+ | **Best Epoch** | 50 |
49
+ | **Current Train Accuracy** | 62.04% |
50
+ | **Current Test Accuracy** | 58.38% |
51
+ | **Current Ξ± (Cantor param)** | 0.4225 |
52
+ | **Total Parameters** | 45,235,067 |
53
+ | **Training Time** | 0:12:52 |
54
 
55
  ### Comparison to State-of-the-Art
56
 
57
  | Model | Accuracy | Status |
58
  |-------|----------|--------|
59
+ | **geo-beatrix (this model)** | **58.38%** | πŸ”„ Training |
60
  | vit-beatrix-dualstream | 66.0% | Vision Transformer + Cross-Entropy |
61
  | CLIP ViT-L/14 (zero-shot) | ~63-65% | 400M image-text pairs |
62
  | CLIP ViT-B/32 (zero-shot) | ~63.5% | Vision Transformer |
63
 
64
+ 🎯 **Current target**: Beat vit-beatrix (66.0%) - Currently -7.62%
65
 
66
  ---
67
 
 
71
  - **Channels**: 64 β†’ 128 β†’ 256 β†’ 512 β†’ 1024
72
  - **Positional Encoding**: Devil's Staircase (Cantor function, 1883)
73
  - **PE Levels**: 20
74
+ - **PE Features/Level**: 10
75
  - **Classification**: Geometric Basin Compatibility (NO cross-entropy)
76
  - **Attention Mechanisms**: NONE
77
 
 
90
  "weight_decay": 0.05,
91
  "warmup_epochs": 10,
92
  "pe_levels": 20,
93
+ "pe_features_per_level": 10,
94
  "dropout": 0.1,
95
+ "upload_every_n_epochs": 50,
96
  "alphamix": {
97
  "enabled": true,
98
  "range": [
 
107
  "loss_function": "Geometric Basin Compatibility",
108
  "cross_entropy": false,
109
  "attention_mechanisms": false,
110
+ "timestamp": "20251009_234907"
111
  }
112
  ```
113
 
 
116
  ## Files Structure
117
 
118
  ```
119
+ β”œβ”€β”€ model.pt (BEST overall model - easy access!)
120
+ β”œβ”€β”€ model.safetensors (BEST overall model - easy access!)
121
+ β”œβ”€β”€ best_model_info.json (which epoch/run this came from)
122
+ β”œβ”€β”€ README.md
123
+ β”œβ”€β”€ weights/geo-beatrix/20251009_234907/
124
+ β”‚ β”œβ”€β”€ model.pt (best from this training run)
125
+ β”‚ β”œβ”€β”€ model.safetensors (best from this training run)
126
+ β”‚ β”œβ”€β”€ config.json
127
+ β”‚ β”œβ”€β”€ training_log.txt
128
+ β”‚ └── checkpoints/
129
+ β”‚ β”œβ”€β”€ checkpoint_epoch_10.safetensors
130
+ β”‚ β”œβ”€β”€ checkpoint_epoch_20.safetensors
131
+ β”‚ └── checkpoint_epoch_30.safetensors
132
+ β”‚ (snapshots every 50 epochs)
133
+ └── runs/geo-beatrix/20251009_234907/
134
+ β”œβ”€β”€ events.out.tfevents.* (TensorBoard logs)
135
+ └── metrics.csv (training metrics)
136
  ```
137
 
138
+ **Note**: The root `model.pt` and `model.safetensors` always contain the best model across all training runs!
139
+
140
  ---
141
 
142
  ## Usage
 
145
  from huggingface_hub import hf_hub_download
146
  import torch
147
 
148
+ # EASIEST: Download BEST overall model from root (recommended!)
149
  from safetensors.torch import load_file
150
  model_path = hf_hub_download(
151
  repo_id="AbstractPhil/geo-beatrix",
152
+ filename="model.safetensors"
153
  )
154
  state_dict = load_file(model_path)
155
  # model.load_state_dict(state_dict)
156
 
157
+ # Check which epoch/run the best model came from
158
+ info_path = hf_hub_download(
159
  repo_id="AbstractPhil/geo-beatrix",
160
+ filename="best_model_info.json"
161
  )
162
+ with open(info_path) as f:
163
+ best_info = json.load(f)
164
+ print(f"Best model: epoch {best_info['epoch']}, {best_info['test_accuracy']:.2f}%")
165
 
166
+ # Or download from specific training run
167
+ model_path = hf_hub_download(
168
  repo_id="AbstractPhil/geo-beatrix",
169
+ filename="weights/geo-beatrix/20251009_234907/model.safetensors"
170
  )
171
 
172
+ # Download specific epoch checkpoint
173
+ epoch_checkpoint = hf_hub_download(
 
174
  repo_id="AbstractPhil/geo-beatrix",
175
+ filename="weights/geo-beatrix/20251009_234907/checkpoints/checkpoint_epoch_100.safetensors"
176
  )
177
  ```
178
 
 
181
  ## Training History
182
 
183
  ### Best Checkpoint
184
+ - Epoch: 50
185
+ - Train Acc: 62.04%
186
+ - Test Acc: 58.38%
187
+ - Alpha: 0.4225
188
+ - Loss: 1.1254
189
 
190
  ### Latest 5 Epochs
191
 
192
+ - **Epoch 46**: Train 59.73%, Test 0.00%, Ξ±=0.4210, Loss=1.1882
193
+ - **Epoch 47**: Train 59.23%, Test 0.00%, Ξ±=0.4176, Loss=1.1413
194
+ - **Epoch 48**: Train 60.73%, Test 0.00%, Ξ±=0.4184, Loss=1.1374
195
+ - **Epoch 49**: Train 60.30%, Test 0.00%, Ξ±=0.4126, Loss=1.1244
196
+ - **Epoch 50**: Train 62.04%, Test 58.38%, Ξ±=0.4225, Loss=1.1254
197
 
198
  ### Training Milestones
199
+ - 🎯 **50% Accuracy** reached at epoch 35
200
+ - πŸ“Š **Ξ± β‰₯ 0.40** reached at epoch 9
 
 
 
 
201
 
202
  ---
203