AbstractPhil commited on
Commit
eaf9bce
·
verified ·
1 Parent(s): 030c272

Update metrics - Run 20251012_132646

Browse files
Files changed (1) hide show
  1. best_model.json +23 -38
best_model.json CHANGED
@@ -1,47 +1,32 @@
1
  {
2
- "model_name": "David-partial_shared-deep_efficiency",
3
- "run_id": "20251012_065325",
4
- "timestamp": "2025-10-12T08:02:12.974571",
5
- "best_val_acc": 83.098,
6
- "best_epoch": 8,
7
- "final_train_acc": 94.51960595301003,
8
- "final_train_loss": 0.5892320626174299,
9
  "scale_accuracies": {
10
- "256": 83.098,
11
- "512": 83.24,
12
- "768": 83.062,
13
- "1024": 83.034,
14
- "1280": 82.922,
15
- "1536": 82.95,
16
- "1792": 82.978,
17
- "2048": 82.968,
18
- "2304": 82.892
19
  },
20
  "architecture": {
21
- "preset": "clip_vit_l14_ultra_deep",
22
- "sharing_mode": "partial_shared",
23
- "fusion_mode": "deep_efficiency",
24
  "scales": [
25
  256,
26
- 512,
27
- 768,
28
- 1024,
29
- 1280,
30
- 1536,
31
- 1792,
32
- 2048,
33
- 2304,
34
- 2560
35
  ],
36
- "feature_dim": 768,
37
  "num_classes": 1000,
38
- "use_belly": true,
39
- "belly_expand": 3.0
40
  },
41
  "training": {
42
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
43
- "model_variant": "clip_vit_l14",
44
- "num_epochs": 10,
45
  "batch_size": 1024,
46
  "learning_rate": 0.001,
47
  "rose_weight": "0.1\u21920.5",
@@ -50,10 +35,10 @@
50
  "scheduler": "cosine_restarts"
51
  },
52
  "files": {
53
- "weights_safetensors": "weights/David-partial_shared-deep_efficiency/20251012_065325/best_model.safetensors",
54
- "weights_pytorch": "weights/David-partial_shared-deep_efficiency/20251012_065325/best_model.pth",
55
- "config": "weights/David-partial_shared-deep_efficiency/20251012_065325/david_config.json",
56
- "training_config": "weights/David-partial_shared-deep_efficiency/20251012_065325/train_config.json",
57
- "tensorboard": "runs/David-partial_shared-deep_efficiency/20251012_065325/"
58
  }
59
  }
 
1
  {
2
+ "model_name": "David-fully_shared-weighted_sum",
3
+ "run_id": "20251012_132646",
4
+ "timestamp": "2025-10-12T13:30:05.861573",
5
+ "best_val_acc": 68.116,
6
+ "best_epoch": 0,
7
+ "final_train_acc": 63.375656725469824,
8
+ "final_train_loss": 5.58824692214259,
9
  "scale_accuracies": {
10
+ "256": 66.996,
11
+ "512": 67.958
 
 
 
 
 
 
 
12
  },
13
  "architecture": {
14
+ "preset": "small_fast",
15
+ "sharing_mode": "fully_shared",
16
+ "fusion_mode": "weighted_sum",
17
  "scales": [
18
  256,
19
+ 512
 
 
 
 
 
 
 
 
20
  ],
21
+ "feature_dim": 512,
22
  "num_classes": 1000,
23
+ "use_belly": false,
24
+ "belly_expand": 2.0
25
  },
26
  "training": {
27
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
28
+ "model_variant": "clip_vit_laion_b32",
29
+ "num_epochs": 20,
30
  "batch_size": 1024,
31
  "learning_rate": 0.001,
32
  "rose_weight": "0.1\u21920.5",
 
35
  "scheduler": "cosine_restarts"
36
  },
37
  "files": {
38
+ "weights_safetensors": "weights/David-fully_shared-weighted_sum/20251012_132646/best_model.safetensors",
39
+ "weights_pytorch": "weights/David-fully_shared-weighted_sum/20251012_132646/best_model.pth",
40
+ "config": "weights/David-fully_shared-weighted_sum/20251012_132646/david_config.json",
41
+ "training_config": "weights/David-fully_shared-weighted_sum/20251012_132646/train_config.json",
42
+ "tensorboard": "runs/David-fully_shared-weighted_sum/20251012_132646/"
43
  }
44
  }