AbstractPhil commited on
Commit
e3cc8c2
·
verified ·
1 Parent(s): f5adac0

Update beatrix-cifar100_20251008_002950 - Epoch 0 - Acc: 0.0469

Browse files
README.md CHANGED
@@ -18,9 +18,20 @@ with geometric simplex features for vision tasks.
18
 
19
  - **Fractal Positional Encoding**: Devil's Staircase multi-scale position embeddings
20
  - **Geometric Simplex Features**: k-simplex vertex computations from Cantor measure
 
21
  - **Adaptive Augmentation**: Progressive augmentation escalation to prevent overfitting
22
  - **Beatrix Formula Suite**: Flow alignment, hierarchical coherence, and multi-scale consistency losses
23
 
 
 
 
 
 
 
 
 
 
 
24
  ### Adaptive Augmentation System
25
 
26
  The trainer includes an intelligent augmentation system that monitors train/validation accuracy gap and progressively enables more augmentation:
@@ -47,41 +58,43 @@ When train accuracy exceeds validation accuracy by 2% or more, the system automa
47
  | beatrix-4simplex-45d | 20251007_231008 | 0.2916 | 85 | `weights/beatrix-4simplex-45d/20251007_231008` | `logs/beatrix-4simplex-45d/20251007_231008` |
48
  | beatrix-cifar100 | 20251007_193112 | 0.2802 | 10 | `weights/beatrix-cifar100/20251007_193112` | `N/A` |
49
  | beatrix-4simplex-45d | 20251008_001147 | 0.1382 | 10 | `weights/beatrix-4simplex-45d/20251008_001147` | `logs/beatrix-4simplex-45d/20251008_001147` |
 
50
 
51
 
52
- ## Latest Updated Model: beatrix-4simplex-45d (Session: 20251008_001147)
53
 
54
  ### Model Details
55
 
56
  - **Architecture**: Vision Transformer with fractal positional encoding
57
  - **Dataset**: CIFAR-100 (100 classes)
58
- - **Embedding Dimension**: 45
59
- - **Depth**: 22 layers
60
  - **Patch Size**: 4x4
61
  - **PE Levels**: 12
62
- - **Simplex Dimension**: 7-simplex
 
63
 
64
  ### Training Details
65
 
66
- - **Training Session**: 20251008_001147
67
- - **Best Accuracy**: 0.1382
68
- - **Epochs Trained**: 10
69
  - **Batch Size**: 512
70
  - **Learning Rate**: 0.0001
71
  - **Adaptive Augmentation**: Enabled
72
 
73
  ### Loss Configuration
74
 
75
- - Task Loss Weight: 1.0
76
- - Flow Alignment Weight: 0.4
77
- - Coherence Weight: 0.2
78
- - Multi-Scale Weight: 0.1
79
 
80
  ### TensorBoard Logs
81
 
82
  Training logs are available in the repository at:
83
  ```
84
- logs/beatrix-4simplex-45d/20251008_001147
85
  ```
86
 
87
  To view them locally:
@@ -90,7 +103,7 @@ To view them locally:
90
  git clone https://huggingface.co/AbstractPhil/vit-beatrix
91
 
92
  # View logs in TensorBoard
93
- tensorboard --logdir vit-beatrix/logs/beatrix-4simplex-45d/20251008_001147
94
  ```
95
 
96
  ## Usage
@@ -135,18 +148,18 @@ with open(manifest_path, 'r') as f:
135
  for key, info in sorted(manifest.items(), key=lambda x: x[1]['accuracy'], reverse=True):
136
  print(f"{info['model_name']} ({info['timestamp']}): {info['accuracy']:.4f}")
137
 
138
- # Download weights for the latest training session of beatrix-4simplex-45d
139
  weights_path = hf_hub_download(
140
  repo_id="AbstractPhil/vit-beatrix",
141
- filename="weights/beatrix-4simplex-45d/20251008_001147/model.safetensors"
142
  )
143
 
144
  # Load model
145
  model = SimplifiedGeometricClassifier(
146
  num_classes=100,
147
  img_size=32,
148
- embed_dim=45,
149
- depth=22
150
  )
151
 
152
  # Load weights
 
18
 
19
  - **Fractal Positional Encoding**: Devil's Staircase multi-scale position embeddings
20
  - **Geometric Simplex Features**: k-simplex vertex computations from Cantor measure
21
+ - **SimplexFactory Initialization**: Pre-initialized simplices with geometrically meaningful shapes (regular/random/uniform)
22
  - **Adaptive Augmentation**: Progressive augmentation escalation to prevent overfitting
23
  - **Beatrix Formula Suite**: Flow alignment, hierarchical coherence, and multi-scale consistency losses
24
 
25
+ ### Simplex Initialization
26
+
27
+ Instead of random initialization, the model uses **SimplexFactory** to create geometrically sound starting configurations:
28
+
29
+ - **Regular** (default): All edges equal length, perfectly balanced symmetric structure
30
+ - **Random**: QR decomposition ensuring affine independence
31
+ - **Uniform**: Hypercube sampling with perturbations
32
+
33
+ Regular simplices provide the most stable and mathematically meaningful initialization, giving the model a better starting point for learning geometric features.
34
+
35
  ### Adaptive Augmentation System
36
 
37
  The trainer includes an intelligent augmentation system that monitors train/validation accuracy gap and progressively enables more augmentation:
 
58
  | beatrix-4simplex-45d | 20251007_231008 | 0.2916 | 85 | `weights/beatrix-4simplex-45d/20251007_231008` | `logs/beatrix-4simplex-45d/20251007_231008` |
59
  | beatrix-cifar100 | 20251007_193112 | 0.2802 | 10 | `weights/beatrix-cifar100/20251007_193112` | `N/A` |
60
  | beatrix-4simplex-45d | 20251008_001147 | 0.1382 | 10 | `weights/beatrix-4simplex-45d/20251008_001147` | `logs/beatrix-4simplex-45d/20251008_001147` |
61
+ | beatrix-cifar100 | 20251008_002950 | 0.0469 | 0 | `weights/beatrix-cifar100/20251008_002950` | `logs/beatrix-cifar100/20251008_002950` |
62
 
63
 
64
+ ## Latest Updated Model: beatrix-cifar100 (Session: 20251008_002950)
65
 
66
  ### Model Details
67
 
68
  - **Architecture**: Vision Transformer with fractal positional encoding
69
  - **Dataset**: CIFAR-100 (100 classes)
70
+ - **Embedding Dimension**: 256
71
+ - **Depth**: 12 layers
72
  - **Patch Size**: 4x4
73
  - **PE Levels**: 12
74
+ - **Simplex Dimension**: 4-simplex
75
+ - **Simplex Initialization**: regular (scale=1.0)
76
 
77
  ### Training Details
78
 
79
+ - **Training Session**: 20251008_002950
80
+ - **Best Accuracy**: 0.0469
81
+ - **Epochs Trained**: 0
82
  - **Batch Size**: 512
83
  - **Learning Rate**: 0.0001
84
  - **Adaptive Augmentation**: Enabled
85
 
86
  ### Loss Configuration
87
 
88
+ - Task Loss Weight: 0.5
89
+ - Flow Alignment Weight: 0.5
90
+ - Coherence Weight: 0.3
91
+ - Multi-Scale Weight: 0.2
92
 
93
  ### TensorBoard Logs
94
 
95
  Training logs are available in the repository at:
96
  ```
97
+ logs/beatrix-cifar100/20251008_002950
98
  ```
99
 
100
  To view them locally:
 
103
  git clone https://huggingface.co/AbstractPhil/vit-beatrix
104
 
105
  # View logs in TensorBoard
106
+ tensorboard --logdir vit-beatrix/logs/beatrix-cifar100/20251008_002950
107
  ```
108
 
109
  ## Usage
 
148
  for key, info in sorted(manifest.items(), key=lambda x: x[1]['accuracy'], reverse=True):
149
  print(f"{info['model_name']} ({info['timestamp']}): {info['accuracy']:.4f}")
150
 
151
+ # Download weights for the latest training session of beatrix-cifar100
152
  weights_path = hf_hub_download(
153
  repo_id="AbstractPhil/vit-beatrix",
154
+ filename="weights/beatrix-cifar100/20251008_002950/model.safetensors"
155
  )
156
 
157
  # Load model
158
  model = SimplifiedGeometricClassifier(
159
  num_classes=100,
160
  img_size=32,
161
+ embed_dim=256,
162
+ depth=12
163
  )
164
 
165
  # Load weights
logs/beatrix-cifar100/20251008_002950/events.out.tfevents.1759883392.f41d735e1a7f.129854.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48662b902788adf15362a46b11f14e37c65abafe3f4ea0caa2f108536da9fcb8
3
+ size 88
manifest.json CHANGED
@@ -79,5 +79,17 @@
79
  "depth": 22,
80
  "batch_size": 512,
81
  "learning_rate": 0.0001
 
 
 
 
 
 
 
 
 
 
 
 
82
  }
83
  }
 
79
  "depth": 22,
80
  "batch_size": 512,
81
  "learning_rate": 0.0001
82
+ },
83
+ "beatrix-cifar100_20251008_002950": {
84
+ "accuracy": 0.046886488981544974,
85
+ "epoch": 0,
86
+ "timestamp": "20251008_002950",
87
+ "model_name": "beatrix-cifar100",
88
+ "path": "weights/beatrix-cifar100/20251008_002950",
89
+ "logs_path": "logs/beatrix-cifar100/20251008_002950",
90
+ "embed_dim": 256,
91
+ "depth": 12,
92
+ "batch_size": 512,
93
+ "learning_rate": 0.0001
94
  }
95
  }
weights/beatrix-cifar100/20251008_002950/config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "num_classes": 100,
3
+ "img_size": 32,
4
+ "patch_size": 4,
5
+ "embed_dim": 256,
6
+ "k_simplex": 4,
7
+ "depth": 12,
8
+ "num_heads": 8,
9
+ "mlp_ratio": 4.0,
10
+ "dropout": 0.0,
11
+ "pe_levels": 12,
12
+ "pe_features_per_level": 2,
13
+ "pe_smooth_tau": 0.25,
14
+ "simplex_feature_weight": 0.5,
15
+ "simplex_init_method": "regular",
16
+ "simplex_init_scale": 1.0,
17
+ "batch_size": 512,
18
+ "num_epochs": 50,
19
+ "learning_rate": 0.0001,
20
+ "weight_decay": 0.005,
21
+ "warmup_epochs": 10,
22
+ "task_loss_weight": 0.5,
23
+ "flow_loss_weight": 0.5,
24
+ "coherence_loss_weight": 0.3,
25
+ "multiscale_loss_weight": 0.2,
26
+ "volume_reg_weight": 0.1,
27
+ "use_adaptive_augmentation": true,
28
+ "overfit_threshold": 0.02,
29
+ "mixup_alpha": 0.2,
30
+ "cutmix_alpha": 1.0,
31
+ "device": "cuda",
32
+ "num_workers": 4,
33
+ "pin_memory": true,
34
+ "save_dir": "./checkpoints",
35
+ "save_every": 10,
36
+ "use_safetensors": true,
37
+ "timestamp_dirs": true,
38
+ "push_to_hub": true,
39
+ "hub_model_id": "AbstractPhil/vit-beatrix",
40
+ "hub_model_name": "beatrix-cifar100",
41
+ "hub_upload_best_only": true,
42
+ "hub_upload_every_n_epochs": 10,
43
+ "use_tensorboard": true,
44
+ "log_dir": "./logs",
45
+ "log_every": 50,
46
+ "train_baseline": false
47
+ }
weights/beatrix-cifar100/20251008_002950/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca4a1d8395406db2e432a0076d17855456d5f8749d1fd6daa2ae9798f4811741
3
+ size 38116128