AbstractPhil commited on
Commit
8f5c228
·
verified ·
1 Parent(s): e8fae8c

Upload weights and configs - Run 20251012_050214

Browse files
weights/best_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cedc682eef6b2abd97c43b3994b107bf28565e59082e0c4844bc193dedca1ab8
3
- size 64195804
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2adddce311fccc3820403cc4fca6539bbd4c15b1f9c498ec363f4afe58ad5787
3
+ size 325845164
weights/best_model_metadata.json CHANGED
The diff for this file is too large to render. See raw diff
 
weights/david_config.json CHANGED
@@ -1,20 +1,26 @@
1
  {
2
- "name": "david_clip_vit_l14",
3
- "uid": "c.david.clip_vit_l14",
4
  "feature_dim": 768,
5
  "num_classes": 1000,
6
  "scales": [
7
- 384,
 
8
  768,
9
  1024,
10
- 1280
 
 
 
 
 
11
  ],
12
  "sharing_mode": "partial_shared",
13
  "fusion_mode": "deep_efficiency",
14
  "use_belly": true,
15
  "belly_expand": 2.0,
16
  "shared_feature_dim": 1024,
17
- "shared_layers": 2,
18
  "shared_dropout": 0.1,
19
  "fusion_temperature": 1.0,
20
  "fusion_dropout": 0.1,
@@ -25,9 +31,15 @@
25
  "attention_dropout": 0.1,
26
  "progressive_training": true,
27
  "scale_warmup_epochs": {
28
- "384": 0,
29
- "768": 1,
30
- "1024": 2,
31
- "1280": 3
 
 
 
 
 
 
32
  }
33
  }
 
1
  {
2
+ "name": "david_clip_vit_l14_deep",
3
+ "uid": "c.david.clip_vit_l14_deep",
4
  "feature_dim": 768,
5
  "num_classes": 1000,
6
  "scales": [
7
+ 256,
8
+ 512,
9
  768,
10
  1024,
11
+ 1280,
12
+ 1536,
13
+ 1792,
14
+ 2048,
15
+ 2304,
16
+ 2560
17
  ],
18
  "sharing_mode": "partial_shared",
19
  "fusion_mode": "deep_efficiency",
20
  "use_belly": true,
21
  "belly_expand": 2.0,
22
  "shared_feature_dim": 1024,
23
+ "shared_layers": 4,
24
  "shared_dropout": 0.1,
25
  "fusion_temperature": 1.0,
26
  "fusion_dropout": 0.1,
 
31
  "attention_dropout": 0.1,
32
  "progressive_training": true,
33
  "scale_warmup_epochs": {
34
+ "256": 0,
35
+ "512": 1,
36
+ "768": 2,
37
+ "1024": 3,
38
+ "1280": 4,
39
+ "1536": 5,
40
+ "1792": 6,
41
+ "2048": 7,
42
+ "2304": 8,
43
+ "2560": 9
44
  }
45
  }
weights/train_config.json CHANGED
@@ -1,26 +1,32 @@
1
  {
2
  "name": "david_training",
3
- "run_id": "20251012_041353",
4
  "dataset_name": "AbstractPhil/imagenet-clip-features-orderly",
5
  "model_variant": "clip_vit_l14",
6
  "num_classes": 1000,
7
- "preset": "clip_vit_l14",
8
  "custom_config_path": null,
9
  "num_classes_override": null,
10
  "use_belly_override": null,
11
  "belly_expand_override": null,
12
  "progressive_training_override": true,
13
  "scale_warmup_epochs_override": {
14
- "384": 0,
15
- "768": 1,
16
- "1024": 2,
17
- "1280": 3
 
 
 
 
 
 
18
  },
19
- "num_epochs": 20,
20
  "batch_size": 1024,
21
  "learning_rate": 0.01,
22
  "weight_decay": 1e-05,
23
- "warmup_epochs": 3,
24
  "use_rose_loss": true,
25
  "rose_initial_weight": 0.1,
26
  "rose_max_weight": 0.5,
 
1
  {
2
  "name": "david_training",
3
+ "run_id": "20251012_050214",
4
  "dataset_name": "AbstractPhil/imagenet-clip-features-orderly",
5
  "model_variant": "clip_vit_l14",
6
  "num_classes": 1000,
7
+ "preset": "clip_vit_l14_deep",
8
  "custom_config_path": null,
9
  "num_classes_override": null,
10
  "use_belly_override": null,
11
  "belly_expand_override": null,
12
  "progressive_training_override": true,
13
  "scale_warmup_epochs_override": {
14
+ "256": 0,
15
+ "512": 1,
16
+ "768": 2,
17
+ "1024": 3,
18
+ "1280": 4,
19
+ "1536": 5,
20
+ "1792": 6,
21
+ "2048": 7,
22
+ "2304": 8,
23
+ "2560": 9
24
  },
25
+ "num_epochs": 10,
26
  "batch_size": 1024,
27
  "learning_rate": 0.01,
28
  "weight_decay": 1e-05,
29
+ "warmup_epochs": 0,
30
  "use_rose_loss": true,
31
  "rose_initial_weight": 0.1,
32
  "rose_max_weight": 0.5,