AbstractPhil commited on
Commit
879f366
·
verified ·
1 Parent(s): 0bb2aa5

Update metrics - Run 20251104_133602

Browse files
Files changed (1) hide show
  1. best_model.json +27 -25
best_model.json CHANGED
@@ -1,37 +1,39 @@
1
  {
2
- "model_name": "David-decoupled-cantor_scale",
3
- "run_id": "20251104_125328",
4
- "timestamp": "2025-11-04T13:32:59.249195",
5
- "best_val_acc": 78.866,
6
- "best_epoch": 4,
7
- "final_train_acc": 87.75288467467551,
8
- "final_train_loss": 1.2123387765693856,
9
  "scale_accuracies": {
10
- "512": 77.274,
11
- "4096": 78.43,
12
- "8192": 78.034,
13
- "16384": 78.344
 
14
  },
15
  "architecture": {
16
- "preset": "clip_vit_b16_cantor_decoupled_massive",
17
- "sharing_mode": "decoupled",
18
  "fusion_mode": "cantor_scale",
19
  "scales": [
20
- 512,
21
- 4096,
22
- 8192,
23
- 16384
 
24
  ],
25
- "feature_dim": 512,
26
  "num_classes": 1000,
27
  "use_belly": true,
28
  "belly_expand": 2.0
29
  },
30
  "training": {
31
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
32
- "model_variant": "clip_vit_b16",
33
  "num_epochs": 5,
34
- "batch_size": 256,
35
  "learning_rate": 0.001,
36
  "rose_weight": "0.1\u21920.5",
37
  "cayley_loss": false,
@@ -39,10 +41,10 @@
39
  "scheduler": "cosine_restarts"
40
  },
41
  "files": {
42
- "weights_safetensors": "weights/David-decoupled-cantor_scale/20251104_125328/best_model_acc78.87.safetensors",
43
- "weights_pytorch": "weights/David-decoupled-cantor_scale/20251104_125328/best_model.pth",
44
- "config": "weights/David-decoupled-cantor_scale/20251104_125328/david_config.json",
45
- "training_config": "weights/David-decoupled-cantor_scale/20251104_125328/train_config.json",
46
- "tensorboard": "runs/David-decoupled-cantor_scale/20251104_125328/"
47
  }
48
  }
 
1
  {
2
+ "model_name": "David-partial_shared-cantor_scale",
3
+ "run_id": "20251104_133602",
4
+ "timestamp": "2025-11-04T13:41:01.936583",
5
+ "best_val_acc": 81.278,
6
+ "best_epoch": 0,
7
+ "final_train_acc": 77.89296789567636,
8
+ "final_train_loss": 2.43972810445883,
9
  "scale_accuracies": {
10
+ "384": 80.394,
11
+ "768": 80.902,
12
+ "1024": 80.752,
13
+ "1280": 80.572,
14
+ "1536": 80.98
15
  },
16
  "architecture": {
17
+ "preset": "clip_vit_l14_cantor",
18
+ "sharing_mode": "partial_shared",
19
  "fusion_mode": "cantor_scale",
20
  "scales": [
21
+ 384,
22
+ 768,
23
+ 1024,
24
+ 1280,
25
+ 1536
26
  ],
27
+ "feature_dim": 768,
28
  "num_classes": 1000,
29
  "use_belly": true,
30
  "belly_expand": 2.0
31
  },
32
  "training": {
33
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
34
+ "model_variant": "clip_vit_l14",
35
  "num_epochs": 5,
36
+ "batch_size": 1024,
37
  "learning_rate": 0.001,
38
  "rose_weight": "0.1\u21920.5",
39
  "cayley_loss": false,
 
41
  "scheduler": "cosine_restarts"
42
  },
43
  "files": {
44
+ "weights_safetensors": "weights/David-partial_shared-cantor_scale/20251104_133602/best_model_acc81.28.safetensors",
45
+ "weights_pytorch": "weights/David-partial_shared-cantor_scale/20251104_133602/best_model.pth",
46
+ "config": "weights/David-partial_shared-cantor_scale/20251104_133602/david_config.json",
47
+ "training_config": "weights/David-partial_shared-cantor_scale/20251104_133602/train_config.json",
48
+ "tensorboard": "runs/David-partial_shared-cantor_scale/20251104_133602/"
49
  }
50
  }