AbstractPhil commited on
Commit
9c604d8
·
verified ·
1 Parent(s): ef7040a

Update metrics - Run 20251012_060013

Browse files
Files changed (1) hide show
  1. best_model.json +16 -32
best_model.json CHANGED
@@ -1,47 +1,31 @@
1
  {
2
- "model_name": "David-partial_shared-deep_efficiency",
3
- "run_id": "20251012_050214",
4
- "timestamp": "2025-10-12T05:26:33.027558",
5
- "best_val_acc": 82.374,
6
- "best_epoch": 9,
7
- "final_train_acc": 90.25771035313898,
8
- "final_train_loss": 0.9337012760650617,
9
  "scale_accuracies": {
10
- "256": 82.374,
11
- "512": 82.728,
12
- "768": 82.702,
13
- "1024": 82.526,
14
- "1280": 82.534,
15
- "1536": 82.482,
16
- "1792": 82.604,
17
- "2048": 82.554,
18
- "2304": 82.604,
19
- "2560": 82.166
20
  },
21
  "architecture": {
22
- "preset": "clip_vit_l14_deep",
23
- "sharing_mode": "partial_shared",
24
- "fusion_mode": "deep_efficiency",
25
  "scales": [
26
  256,
27
- 512,
28
- 768,
29
- 1024,
30
- 1280,
31
- 1536,
32
- 1792,
33
- 2048,
34
- 2304,
35
- 2560
36
  ],
37
- "feature_dim": 768,
38
  "num_classes": 1000,
39
- "use_belly": true,
40
  "belly_expand": 2.0
41
  },
42
  "training": {
43
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
44
- "model_variant": "clip_vit_l14",
45
  "num_epochs": 10,
46
  "batch_size": 1024,
47
  "learning_rate": 0.01,
 
1
  {
2
+ "model_name": "David-fully_shared-weighted_sum",
3
+ "run_id": "20251012_060013",
4
+ "timestamp": "2025-10-12T06:02:00.550898",
5
+ "best_val_acc": 72.376,
6
+ "best_epoch": 0,
7
+ "final_train_acc": 66.84764749638416,
8
+ "final_train_loss": 4.327203201028866,
9
  "scale_accuracies": {
10
+ "256": 71.828,
11
+ "512": 72.234
 
 
 
 
 
 
 
 
12
  },
13
  "architecture": {
14
+ "preset": "small_fast",
15
+ "sharing_mode": "fully_shared",
16
+ "fusion_mode": "weighted_sum",
17
  "scales": [
18
  256,
19
+ 512
 
 
 
 
 
 
 
 
20
  ],
21
+ "feature_dim": 512,
22
  "num_classes": 1000,
23
+ "use_belly": false,
24
  "belly_expand": 2.0
25
  },
26
  "training": {
27
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
28
+ "model_variant": "clip_vit_b16",
29
  "num_epochs": 10,
30
  "batch_size": 1024,
31
  "learning_rate": 0.01,