AbstractPhil commited on
Commit
a841477
·
verified ·
1 Parent(s): e06f060

Update metrics - Run 20251012_065325

Browse files
Files changed (1) hide show
  1. best_model.json +25 -22
best_model.json CHANGED
@@ -1,36 +1,39 @@
1
  {
2
- "model_name": "David-partial_shared-hierarchical_tree",
3
  "run_id": "20251012_065325",
4
- "timestamp": "2025-10-12T07:35:40.795003",
5
- "best_val_acc": 75.414,
6
- "best_epoch": 9,
7
- "final_train_acc": 87.91102174814056,
8
- "final_train_loss": 1.3676543517615467,
9
  "scale_accuracies": {
10
- "256": 74.794,
11
- "512": 75.394,
12
- "768": 75.398,
13
- "1024": 73.42
14
  },
15
  "architecture": {
16
- "preset": "balanced",
17
  "sharing_mode": "partial_shared",
18
- "fusion_mode": "hierarchical_tree",
19
  "scales": [
20
  256,
21
  512,
22
  768,
23
- 1024
 
 
 
 
 
 
24
  ],
25
- "feature_dim": 512,
26
  "num_classes": 1000,
27
  "use_belly": true,
28
- "belly_expand": 2.0
29
  },
30
  "training": {
31
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
32
- "model_variant": "clip_vit_laion_b32",
33
- "num_epochs": 20,
34
  "batch_size": 1024,
35
  "learning_rate": 0.001,
36
  "rose_weight": "0.1\u21920.5",
@@ -39,10 +42,10 @@
39
  "scheduler": "cosine_restarts"
40
  },
41
  "files": {
42
- "weights_safetensors": "weights/David-partial_shared-hierarchical_tree/20251012_065325/best_model.safetensors",
43
- "weights_pytorch": "weights/David-partial_shared-hierarchical_tree/20251012_065325/best_model.pth",
44
- "config": "weights/David-partial_shared-hierarchical_tree/20251012_065325/david_config.json",
45
- "training_config": "weights/David-partial_shared-hierarchical_tree/20251012_065325/train_config.json",
46
- "tensorboard": "runs/David-partial_shared-hierarchical_tree/20251012_065325/"
47
  }
48
  }
 
1
  {
2
+ "model_name": "David-partial_shared-deep_efficiency",
3
  "run_id": "20251012_065325",
4
+ "timestamp": "2025-10-12T07:38:25.221146",
5
+ "best_val_acc": 81.158,
6
+ "best_epoch": 0,
7
+ "final_train_acc": 78.10129358623817,
8
+ "final_train_loss": 1.346226876060041,
9
  "scale_accuracies": {
10
+ "256": 81.158
 
 
 
11
  },
12
  "architecture": {
13
+ "preset": "clip_vit_l14_ultra_deep",
14
  "sharing_mode": "partial_shared",
15
+ "fusion_mode": "deep_efficiency",
16
  "scales": [
17
  256,
18
  512,
19
  768,
20
+ 1024,
21
+ 1280,
22
+ 1536,
23
+ 1792,
24
+ 2048,
25
+ 2304,
26
+ 2560
27
  ],
28
+ "feature_dim": 768,
29
  "num_classes": 1000,
30
  "use_belly": true,
31
+ "belly_expand": 3.0
32
  },
33
  "training": {
34
  "dataset": "AbstractPhil/imagenet-clip-features-orderly",
35
+ "model_variant": "clip_vit_l14",
36
+ "num_epochs": 10,
37
  "batch_size": 1024,
38
  "learning_rate": 0.001,
39
  "rose_weight": "0.1\u21920.5",
 
42
  "scheduler": "cosine_restarts"
43
  },
44
  "files": {
45
+ "weights_safetensors": "weights/David-partial_shared-deep_efficiency/20251012_065325/best_model.safetensors",
46
+ "weights_pytorch": "weights/David-partial_shared-deep_efficiency/20251012_065325/best_model.pth",
47
+ "config": "weights/David-partial_shared-deep_efficiency/20251012_065325/david_config.json",
48
+ "training_config": "weights/David-partial_shared-deep_efficiency/20251012_065325/train_config.json",
49
+ "tensorboard": "runs/David-partial_shared-deep_efficiency/20251012_065325/"
50
  }
51
  }