tanganke commited on
Commit
e16bd2a
·
verified ·
1 Parent(s): b60ffc5

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. resnet152-cifar10-batch_size_128_lr_0.001_training_data_ratio_0.5-2000/model.safetensors +3 -0
  2. resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_0.5-4000/README.md +76 -0
  3. resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_0.5-4000/config.json +63 -0
  4. resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_0.5-4000/preprocessor_config.json +22 -0
  5. resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_1.0-3000/model.safetensors +3 -0
  6. resnet152-cifar10-batch_size_256_lr_0.005_training_data_ratio_0.5-2000/model.safetensors +3 -0
  7. resnet152-cifar10-batch_size_256_lr_0.01_training_data_ratio_1.0-3000/README.md +76 -0
  8. resnet152-cifar10-batch_size_256_lr_0.01_training_data_ratio_1.0-3000/config.json +63 -0
  9. resnet152-cifar10-batch_size_256_lr_0.01_training_data_ratio_1.0-3000/preprocessor_config.json +22 -0
  10. resnet152-cifar10-batch_size_64_lr_0.001_training_data_ratio_1.0-3000/README.md +76 -0
  11. resnet152-cifar10-batch_size_64_lr_0.001_training_data_ratio_1.0-3000/config.json +63 -0
  12. resnet152-cifar10-batch_size_64_lr_0.001_training_data_ratio_1.0-3000/preprocessor_config.json +22 -0
  13. resnet152-cifar10-batch_size_64_lr_0.01_training_data_ratio_0.8-3000/model.safetensors +3 -0
  14. resnet152-cifar100-batch_size_128_lr_0.01_training_data_ratio_0.5-1000/model.safetensors +3 -0
  15. resnet152-cifar100-batch_size_256_lr_0.005_training_data_ratio_0.5-1000/model.safetensors +3 -0
  16. resnet152-cifar100-batch_size_256_lr_0.005_training_data_ratio_0.8-2000/model.safetensors +3 -0
  17. resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-2000/README.md +76 -0
  18. resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-2000/config.json +243 -0
  19. resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-2000/preprocessor_config.json +22 -0
  20. resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-4000/model.safetensors +3 -0
  21. resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_0.5-3000/README.md +76 -0
  22. resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_0.5-3000/config.json +243 -0
  23. resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_0.5-3000/preprocessor_config.json +22 -0
  24. resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_1.0-2000/README.md +76 -0
  25. resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_1.0-2000/config.json +243 -0
  26. resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_1.0-2000/preprocessor_config.json +22 -0
  27. resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_0.5-3000/model.safetensors +3 -0
  28. resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_0.8-4000/model.safetensors +3 -0
  29. resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_1.0-1000/README.md +76 -0
  30. resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_1.0-1000/config.json +243 -0
  31. resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_1.0-1000/preprocessor_config.json +22 -0
  32. resnet152-dtd-batch_size_128_lr_0.001_training_data_ratio_0.8-4000/model.safetensors +3 -0
  33. resnet152-dtd-batch_size_128_lr_0.005_training_data_ratio_0.8-2000/model.safetensors +3 -0
  34. resnet152-dtd-batch_size_128_lr_0.005_training_data_ratio_0.8-4000/model.safetensors +3 -0
  35. resnet152-dtd-batch_size_128_lr_0.01_training_data_ratio_0.8-4000/model.safetensors +3 -0
  36. resnet152-dtd-batch_size_256_lr_0.005_training_data_ratio_1.0-3000/model.safetensors +3 -0
  37. resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_0.8-2000/README.md +76 -0
  38. resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_0.8-2000/config.json +137 -0
  39. resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_0.8-2000/preprocessor_config.json +22 -0
  40. resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_1.0-2000/model.safetensors +3 -0
  41. resnet152-dtd-batch_size_64_lr_0.005_training_data_ratio_0.8-1000/model.safetensors +3 -0
  42. resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_0.5-1000/README.md +76 -0
  43. resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_0.5-1000/config.json +137 -0
  44. resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_0.5-1000/preprocessor_config.json +22 -0
  45. resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_1.0-1000/README.md +76 -0
  46. resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_1.0-1000/config.json +137 -0
  47. resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_1.0-1000/preprocessor_config.json +22 -0
  48. resnet152-emnist_letters-batch_size_128_lr_0.001_training_data_ratio_0.8-4000/model.safetensors +3 -0
  49. resnet152-emnist_letters-batch_size_128_lr_0.001_training_data_ratio_1.0-4000/README.md +76 -0
  50. resnet152-emnist_letters-batch_size_128_lr_0.001_training_data_ratio_1.0-4000/config.json +95 -0
resnet152-cifar10-batch_size_128_lr_0.001_training_data_ratio_0.5-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f35a1570bffc474f337e313720d683f2ebcf056bce90092b4cd5b29a8991414c
3
+ size 233387400
resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_0.5-4000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar10.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 128
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.001
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 0.5
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar10
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar10:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar10
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar10:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar10
74
+ split: test
75
+ ```
76
+
resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_0.5-4000/config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "airplane",
24
+ "1": "automobile",
25
+ "2": "bird",
26
+ "3": "cat",
27
+ "4": "deer",
28
+ "5": "dog",
29
+ "6": "frog",
30
+ "7": "horse",
31
+ "8": "ship",
32
+ "9": "truck"
33
+ },
34
+ "label2id": {
35
+ "airplane": 0,
36
+ "automobile": 1,
37
+ "bird": 2,
38
+ "cat": 3,
39
+ "deer": 4,
40
+ "dog": 5,
41
+ "frog": 6,
42
+ "horse": 7,
43
+ "ship": 8,
44
+ "truck": 9
45
+ },
46
+ "layer_type": "bottleneck",
47
+ "model_type": "resnet",
48
+ "num_channels": 3,
49
+ "out_features": [
50
+ "stage4"
51
+ ],
52
+ "out_indices": [
53
+ 4
54
+ ],
55
+ "stage_names": [
56
+ "stem",
57
+ "stage1",
58
+ "stage2",
59
+ "stage3",
60
+ "stage4"
61
+ ],
62
+ "transformers_version": "4.56.1"
63
+ }
resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_0.5-4000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-cifar10-batch_size_256_lr_0.001_training_data_ratio_1.0-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb318a35e3b325f2f76d1212be89aefd23175f9286f4664209b6dba9408e33e
3
+ size 233387400
resnet152-cifar10-batch_size_256_lr_0.005_training_data_ratio_0.5-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1ed2e154399932fe39ab7ef4efbc61f214b0b101f1b2809ccea1ba02e5a4602
3
+ size 233387400
resnet152-cifar10-batch_size_256_lr_0.01_training_data_ratio_1.0-3000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar10.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 128
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.01
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 1.0
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar10
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar10:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar10
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar10:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar10
74
+ split: test
75
+ ```
76
+
resnet152-cifar10-batch_size_256_lr_0.01_training_data_ratio_1.0-3000/config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "airplane",
24
+ "1": "automobile",
25
+ "2": "bird",
26
+ "3": "cat",
27
+ "4": "deer",
28
+ "5": "dog",
29
+ "6": "frog",
30
+ "7": "horse",
31
+ "8": "ship",
32
+ "9": "truck"
33
+ },
34
+ "label2id": {
35
+ "airplane": 0,
36
+ "automobile": 1,
37
+ "bird": 2,
38
+ "cat": 3,
39
+ "deer": 4,
40
+ "dog": 5,
41
+ "frog": 6,
42
+ "horse": 7,
43
+ "ship": 8,
44
+ "truck": 9
45
+ },
46
+ "layer_type": "bottleneck",
47
+ "model_type": "resnet",
48
+ "num_channels": 3,
49
+ "out_features": [
50
+ "stage4"
51
+ ],
52
+ "out_indices": [
53
+ 4
54
+ ],
55
+ "stage_names": [
56
+ "stem",
57
+ "stage1",
58
+ "stage2",
59
+ "stage3",
60
+ "stage4"
61
+ ],
62
+ "transformers_version": "4.56.1"
63
+ }
resnet152-cifar10-batch_size_256_lr_0.01_training_data_ratio_1.0-3000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-cifar10-batch_size_64_lr_0.001_training_data_ratio_1.0-3000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar10.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 64
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.001
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 1.0
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar10
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar10:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar10
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar10:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar10
74
+ split: test
75
+ ```
76
+
resnet152-cifar10-batch_size_64_lr_0.001_training_data_ratio_1.0-3000/config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "airplane",
24
+ "1": "automobile",
25
+ "2": "bird",
26
+ "3": "cat",
27
+ "4": "deer",
28
+ "5": "dog",
29
+ "6": "frog",
30
+ "7": "horse",
31
+ "8": "ship",
32
+ "9": "truck"
33
+ },
34
+ "label2id": {
35
+ "airplane": 0,
36
+ "automobile": 1,
37
+ "bird": 2,
38
+ "cat": 3,
39
+ "deer": 4,
40
+ "dog": 5,
41
+ "frog": 6,
42
+ "horse": 7,
43
+ "ship": 8,
44
+ "truck": 9
45
+ },
46
+ "layer_type": "bottleneck",
47
+ "model_type": "resnet",
48
+ "num_channels": 3,
49
+ "out_features": [
50
+ "stage4"
51
+ ],
52
+ "out_indices": [
53
+ 4
54
+ ],
55
+ "stage_names": [
56
+ "stem",
57
+ "stage1",
58
+ "stage2",
59
+ "stage3",
60
+ "stage4"
61
+ ],
62
+ "transformers_version": "4.56.1"
63
+ }
resnet152-cifar10-batch_size_64_lr_0.001_training_data_ratio_1.0-3000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-cifar10-batch_size_64_lr_0.01_training_data_ratio_0.8-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3894476e72af4c97d37cb1d315111117fe58fe949095ac764b5754d3261e61e
3
+ size 233387400
resnet152-cifar100-batch_size_128_lr_0.01_training_data_ratio_0.5-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98e2037b6c8811dd62a919bd44b8ef84c607eb975b960bcefc6f2ce4fc965a41
3
+ size 234125152
resnet152-cifar100-batch_size_256_lr_0.005_training_data_ratio_0.5-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4007645ebbce0a0af1925d6f77587d0a10fbaf11431bc42ce5537506d8ca95a1
3
+ size 234125152
resnet152-cifar100-batch_size_256_lr_0.005_training_data_ratio_0.8-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e71d30b530ef6249015b8ea50458313828a9c4476ba402f60eb5343114b0979
3
+ size 234125152
resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-2000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar100.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 128
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.01
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 0.5
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar100
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar100:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar100
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar100:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar100
74
+ split: test
75
+ ```
76
+
resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-2000/config.json ADDED
@@ -0,0 +1,243 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "apple",
24
+ "1": "aquarium_fish",
25
+ "2": "baby",
26
+ "3": "bear",
27
+ "4": "beaver",
28
+ "5": "bed",
29
+ "6": "bee",
30
+ "7": "beetle",
31
+ "8": "bicycle",
32
+ "9": "bottle",
33
+ "10": "bowl",
34
+ "11": "boy",
35
+ "12": "bridge",
36
+ "13": "bus",
37
+ "14": "butterfly",
38
+ "15": "camel",
39
+ "16": "can",
40
+ "17": "castle",
41
+ "18": "caterpillar",
42
+ "19": "cattle",
43
+ "20": "chair",
44
+ "21": "chimpanzee",
45
+ "22": "clock",
46
+ "23": "cloud",
47
+ "24": "cockroach",
48
+ "25": "couch",
49
+ "26": "cra",
50
+ "27": "crocodile",
51
+ "28": "cup",
52
+ "29": "dinosaur",
53
+ "30": "dolphin",
54
+ "31": "elephant",
55
+ "32": "flatfish",
56
+ "33": "forest",
57
+ "34": "fox",
58
+ "35": "girl",
59
+ "36": "hamster",
60
+ "37": "house",
61
+ "38": "kangaroo",
62
+ "39": "keyboard",
63
+ "40": "lamp",
64
+ "41": "lawn_mower",
65
+ "42": "leopard",
66
+ "43": "lion",
67
+ "44": "lizard",
68
+ "45": "lobster",
69
+ "46": "man",
70
+ "47": "maple_tree",
71
+ "48": "motorcycle",
72
+ "49": "mountain",
73
+ "50": "mouse",
74
+ "51": "mushroom",
75
+ "52": "oak_tree",
76
+ "53": "orange",
77
+ "54": "orchid",
78
+ "55": "otter",
79
+ "56": "palm_tree",
80
+ "57": "pear",
81
+ "58": "pickup_truck",
82
+ "59": "pine_tree",
83
+ "60": "plain",
84
+ "61": "plate",
85
+ "62": "poppy",
86
+ "63": "porcupine",
87
+ "64": "possum",
88
+ "65": "rabbit",
89
+ "66": "raccoon",
90
+ "67": "ray",
91
+ "68": "road",
92
+ "69": "rocket",
93
+ "70": "rose",
94
+ "71": "sea",
95
+ "72": "seal",
96
+ "73": "shark",
97
+ "74": "shrew",
98
+ "75": "skunk",
99
+ "76": "skyscraper",
100
+ "77": "snail",
101
+ "78": "snake",
102
+ "79": "spider",
103
+ "80": "squirrel",
104
+ "81": "streetcar",
105
+ "82": "sunflower",
106
+ "83": "sweet_pepper",
107
+ "84": "table",
108
+ "85": "tank",
109
+ "86": "telephone",
110
+ "87": "television",
111
+ "88": "tiger",
112
+ "89": "tractor",
113
+ "90": "train",
114
+ "91": "trout",
115
+ "92": "tulip",
116
+ "93": "turtle",
117
+ "94": "wardrobe",
118
+ "95": "whale",
119
+ "96": "willow_tree",
120
+ "97": "wolf",
121
+ "98": "woman",
122
+ "99": "worm"
123
+ },
124
+ "label2id": {
125
+ "apple": 0,
126
+ "aquarium_fish": 1,
127
+ "baby": 2,
128
+ "bear": 3,
129
+ "beaver": 4,
130
+ "bed": 5,
131
+ "bee": 6,
132
+ "beetle": 7,
133
+ "bicycle": 8,
134
+ "bottle": 9,
135
+ "bowl": 10,
136
+ "boy": 11,
137
+ "bridge": 12,
138
+ "bus": 13,
139
+ "butterfly": 14,
140
+ "camel": 15,
141
+ "can": 16,
142
+ "castle": 17,
143
+ "caterpillar": 18,
144
+ "cattle": 19,
145
+ "chair": 20,
146
+ "chimpanzee": 21,
147
+ "clock": 22,
148
+ "cloud": 23,
149
+ "cockroach": 24,
150
+ "couch": 25,
151
+ "cra": 26,
152
+ "crocodile": 27,
153
+ "cup": 28,
154
+ "dinosaur": 29,
155
+ "dolphin": 30,
156
+ "elephant": 31,
157
+ "flatfish": 32,
158
+ "forest": 33,
159
+ "fox": 34,
160
+ "girl": 35,
161
+ "hamster": 36,
162
+ "house": 37,
163
+ "kangaroo": 38,
164
+ "keyboard": 39,
165
+ "lamp": 40,
166
+ "lawn_mower": 41,
167
+ "leopard": 42,
168
+ "lion": 43,
169
+ "lizard": 44,
170
+ "lobster": 45,
171
+ "man": 46,
172
+ "maple_tree": 47,
173
+ "motorcycle": 48,
174
+ "mountain": 49,
175
+ "mouse": 50,
176
+ "mushroom": 51,
177
+ "oak_tree": 52,
178
+ "orange": 53,
179
+ "orchid": 54,
180
+ "otter": 55,
181
+ "palm_tree": 56,
182
+ "pear": 57,
183
+ "pickup_truck": 58,
184
+ "pine_tree": 59,
185
+ "plain": 60,
186
+ "plate": 61,
187
+ "poppy": 62,
188
+ "porcupine": 63,
189
+ "possum": 64,
190
+ "rabbit": 65,
191
+ "raccoon": 66,
192
+ "ray": 67,
193
+ "road": 68,
194
+ "rocket": 69,
195
+ "rose": 70,
196
+ "sea": 71,
197
+ "seal": 72,
198
+ "shark": 73,
199
+ "shrew": 74,
200
+ "skunk": 75,
201
+ "skyscraper": 76,
202
+ "snail": 77,
203
+ "snake": 78,
204
+ "spider": 79,
205
+ "squirrel": 80,
206
+ "streetcar": 81,
207
+ "sunflower": 82,
208
+ "sweet_pepper": 83,
209
+ "table": 84,
210
+ "tank": 85,
211
+ "telephone": 86,
212
+ "television": 87,
213
+ "tiger": 88,
214
+ "tractor": 89,
215
+ "train": 90,
216
+ "trout": 91,
217
+ "tulip": 92,
218
+ "turtle": 93,
219
+ "wardrobe": 94,
220
+ "whale": 95,
221
+ "willow_tree": 96,
222
+ "wolf": 97,
223
+ "woman": 98,
224
+ "worm": 99
225
+ },
226
+ "layer_type": "bottleneck",
227
+ "model_type": "resnet",
228
+ "num_channels": 3,
229
+ "out_features": [
230
+ "stage4"
231
+ ],
232
+ "out_indices": [
233
+ 4
234
+ ],
235
+ "stage_names": [
236
+ "stem",
237
+ "stage1",
238
+ "stage2",
239
+ "stage3",
240
+ "stage4"
241
+ ],
242
+ "transformers_version": "4.56.1"
243
+ }
resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-2000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-cifar100-batch_size_256_lr_0.01_training_data_ratio_0.5-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a3f5f880f0621374a0052c1704175c34e33562dd9bf4d9ca220c5d50000a4fa
3
+ size 234125152
resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_0.5-3000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar100.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 64
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.001
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 0.5
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar100
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar100:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar100
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar100:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar100
74
+ split: test
75
+ ```
76
+
resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_0.5-3000/config.json ADDED
@@ -0,0 +1,243 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "apple",
24
+ "1": "aquarium_fish",
25
+ "2": "baby",
26
+ "3": "bear",
27
+ "4": "beaver",
28
+ "5": "bed",
29
+ "6": "bee",
30
+ "7": "beetle",
31
+ "8": "bicycle",
32
+ "9": "bottle",
33
+ "10": "bowl",
34
+ "11": "boy",
35
+ "12": "bridge",
36
+ "13": "bus",
37
+ "14": "butterfly",
38
+ "15": "camel",
39
+ "16": "can",
40
+ "17": "castle",
41
+ "18": "caterpillar",
42
+ "19": "cattle",
43
+ "20": "chair",
44
+ "21": "chimpanzee",
45
+ "22": "clock",
46
+ "23": "cloud",
47
+ "24": "cockroach",
48
+ "25": "couch",
49
+ "26": "cra",
50
+ "27": "crocodile",
51
+ "28": "cup",
52
+ "29": "dinosaur",
53
+ "30": "dolphin",
54
+ "31": "elephant",
55
+ "32": "flatfish",
56
+ "33": "forest",
57
+ "34": "fox",
58
+ "35": "girl",
59
+ "36": "hamster",
60
+ "37": "house",
61
+ "38": "kangaroo",
62
+ "39": "keyboard",
63
+ "40": "lamp",
64
+ "41": "lawn_mower",
65
+ "42": "leopard",
66
+ "43": "lion",
67
+ "44": "lizard",
68
+ "45": "lobster",
69
+ "46": "man",
70
+ "47": "maple_tree",
71
+ "48": "motorcycle",
72
+ "49": "mountain",
73
+ "50": "mouse",
74
+ "51": "mushroom",
75
+ "52": "oak_tree",
76
+ "53": "orange",
77
+ "54": "orchid",
78
+ "55": "otter",
79
+ "56": "palm_tree",
80
+ "57": "pear",
81
+ "58": "pickup_truck",
82
+ "59": "pine_tree",
83
+ "60": "plain",
84
+ "61": "plate",
85
+ "62": "poppy",
86
+ "63": "porcupine",
87
+ "64": "possum",
88
+ "65": "rabbit",
89
+ "66": "raccoon",
90
+ "67": "ray",
91
+ "68": "road",
92
+ "69": "rocket",
93
+ "70": "rose",
94
+ "71": "sea",
95
+ "72": "seal",
96
+ "73": "shark",
97
+ "74": "shrew",
98
+ "75": "skunk",
99
+ "76": "skyscraper",
100
+ "77": "snail",
101
+ "78": "snake",
102
+ "79": "spider",
103
+ "80": "squirrel",
104
+ "81": "streetcar",
105
+ "82": "sunflower",
106
+ "83": "sweet_pepper",
107
+ "84": "table",
108
+ "85": "tank",
109
+ "86": "telephone",
110
+ "87": "television",
111
+ "88": "tiger",
112
+ "89": "tractor",
113
+ "90": "train",
114
+ "91": "trout",
115
+ "92": "tulip",
116
+ "93": "turtle",
117
+ "94": "wardrobe",
118
+ "95": "whale",
119
+ "96": "willow_tree",
120
+ "97": "wolf",
121
+ "98": "woman",
122
+ "99": "worm"
123
+ },
124
+ "label2id": {
125
+ "apple": 0,
126
+ "aquarium_fish": 1,
127
+ "baby": 2,
128
+ "bear": 3,
129
+ "beaver": 4,
130
+ "bed": 5,
131
+ "bee": 6,
132
+ "beetle": 7,
133
+ "bicycle": 8,
134
+ "bottle": 9,
135
+ "bowl": 10,
136
+ "boy": 11,
137
+ "bridge": 12,
138
+ "bus": 13,
139
+ "butterfly": 14,
140
+ "camel": 15,
141
+ "can": 16,
142
+ "castle": 17,
143
+ "caterpillar": 18,
144
+ "cattle": 19,
145
+ "chair": 20,
146
+ "chimpanzee": 21,
147
+ "clock": 22,
148
+ "cloud": 23,
149
+ "cockroach": 24,
150
+ "couch": 25,
151
+ "cra": 26,
152
+ "crocodile": 27,
153
+ "cup": 28,
154
+ "dinosaur": 29,
155
+ "dolphin": 30,
156
+ "elephant": 31,
157
+ "flatfish": 32,
158
+ "forest": 33,
159
+ "fox": 34,
160
+ "girl": 35,
161
+ "hamster": 36,
162
+ "house": 37,
163
+ "kangaroo": 38,
164
+ "keyboard": 39,
165
+ "lamp": 40,
166
+ "lawn_mower": 41,
167
+ "leopard": 42,
168
+ "lion": 43,
169
+ "lizard": 44,
170
+ "lobster": 45,
171
+ "man": 46,
172
+ "maple_tree": 47,
173
+ "motorcycle": 48,
174
+ "mountain": 49,
175
+ "mouse": 50,
176
+ "mushroom": 51,
177
+ "oak_tree": 52,
178
+ "orange": 53,
179
+ "orchid": 54,
180
+ "otter": 55,
181
+ "palm_tree": 56,
182
+ "pear": 57,
183
+ "pickup_truck": 58,
184
+ "pine_tree": 59,
185
+ "plain": 60,
186
+ "plate": 61,
187
+ "poppy": 62,
188
+ "porcupine": 63,
189
+ "possum": 64,
190
+ "rabbit": 65,
191
+ "raccoon": 66,
192
+ "ray": 67,
193
+ "road": 68,
194
+ "rocket": 69,
195
+ "rose": 70,
196
+ "sea": 71,
197
+ "seal": 72,
198
+ "shark": 73,
199
+ "shrew": 74,
200
+ "skunk": 75,
201
+ "skyscraper": 76,
202
+ "snail": 77,
203
+ "snake": 78,
204
+ "spider": 79,
205
+ "squirrel": 80,
206
+ "streetcar": 81,
207
+ "sunflower": 82,
208
+ "sweet_pepper": 83,
209
+ "table": 84,
210
+ "tank": 85,
211
+ "telephone": 86,
212
+ "television": 87,
213
+ "tiger": 88,
214
+ "tractor": 89,
215
+ "train": 90,
216
+ "trout": 91,
217
+ "tulip": 92,
218
+ "turtle": 93,
219
+ "wardrobe": 94,
220
+ "whale": 95,
221
+ "willow_tree": 96,
222
+ "wolf": 97,
223
+ "woman": 98,
224
+ "worm": 99
225
+ },
226
+ "layer_type": "bottleneck",
227
+ "model_type": "resnet",
228
+ "num_channels": 3,
229
+ "out_features": [
230
+ "stage4"
231
+ ],
232
+ "out_indices": [
233
+ 4
234
+ ],
235
+ "stage_names": [
236
+ "stem",
237
+ "stage1",
238
+ "stage2",
239
+ "stage3",
240
+ "stage4"
241
+ ],
242
+ "transformers_version": "4.56.1"
243
+ }
resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_0.5-3000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_1.0-2000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar100.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 64
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.001
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 1.0
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar100
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar100:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar100
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar100:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar100
74
+ split: test
75
+ ```
76
+
resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_1.0-2000/config.json ADDED
@@ -0,0 +1,243 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "apple",
24
+ "1": "aquarium_fish",
25
+ "2": "baby",
26
+ "3": "bear",
27
+ "4": "beaver",
28
+ "5": "bed",
29
+ "6": "bee",
30
+ "7": "beetle",
31
+ "8": "bicycle",
32
+ "9": "bottle",
33
+ "10": "bowl",
34
+ "11": "boy",
35
+ "12": "bridge",
36
+ "13": "bus",
37
+ "14": "butterfly",
38
+ "15": "camel",
39
+ "16": "can",
40
+ "17": "castle",
41
+ "18": "caterpillar",
42
+ "19": "cattle",
43
+ "20": "chair",
44
+ "21": "chimpanzee",
45
+ "22": "clock",
46
+ "23": "cloud",
47
+ "24": "cockroach",
48
+ "25": "couch",
49
+ "26": "cra",
50
+ "27": "crocodile",
51
+ "28": "cup",
52
+ "29": "dinosaur",
53
+ "30": "dolphin",
54
+ "31": "elephant",
55
+ "32": "flatfish",
56
+ "33": "forest",
57
+ "34": "fox",
58
+ "35": "girl",
59
+ "36": "hamster",
60
+ "37": "house",
61
+ "38": "kangaroo",
62
+ "39": "keyboard",
63
+ "40": "lamp",
64
+ "41": "lawn_mower",
65
+ "42": "leopard",
66
+ "43": "lion",
67
+ "44": "lizard",
68
+ "45": "lobster",
69
+ "46": "man",
70
+ "47": "maple_tree",
71
+ "48": "motorcycle",
72
+ "49": "mountain",
73
+ "50": "mouse",
74
+ "51": "mushroom",
75
+ "52": "oak_tree",
76
+ "53": "orange",
77
+ "54": "orchid",
78
+ "55": "otter",
79
+ "56": "palm_tree",
80
+ "57": "pear",
81
+ "58": "pickup_truck",
82
+ "59": "pine_tree",
83
+ "60": "plain",
84
+ "61": "plate",
85
+ "62": "poppy",
86
+ "63": "porcupine",
87
+ "64": "possum",
88
+ "65": "rabbit",
89
+ "66": "raccoon",
90
+ "67": "ray",
91
+ "68": "road",
92
+ "69": "rocket",
93
+ "70": "rose",
94
+ "71": "sea",
95
+ "72": "seal",
96
+ "73": "shark",
97
+ "74": "shrew",
98
+ "75": "skunk",
99
+ "76": "skyscraper",
100
+ "77": "snail",
101
+ "78": "snake",
102
+ "79": "spider",
103
+ "80": "squirrel",
104
+ "81": "streetcar",
105
+ "82": "sunflower",
106
+ "83": "sweet_pepper",
107
+ "84": "table",
108
+ "85": "tank",
109
+ "86": "telephone",
110
+ "87": "television",
111
+ "88": "tiger",
112
+ "89": "tractor",
113
+ "90": "train",
114
+ "91": "trout",
115
+ "92": "tulip",
116
+ "93": "turtle",
117
+ "94": "wardrobe",
118
+ "95": "whale",
119
+ "96": "willow_tree",
120
+ "97": "wolf",
121
+ "98": "woman",
122
+ "99": "worm"
123
+ },
124
+ "label2id": {
125
+ "apple": 0,
126
+ "aquarium_fish": 1,
127
+ "baby": 2,
128
+ "bear": 3,
129
+ "beaver": 4,
130
+ "bed": 5,
131
+ "bee": 6,
132
+ "beetle": 7,
133
+ "bicycle": 8,
134
+ "bottle": 9,
135
+ "bowl": 10,
136
+ "boy": 11,
137
+ "bridge": 12,
138
+ "bus": 13,
139
+ "butterfly": 14,
140
+ "camel": 15,
141
+ "can": 16,
142
+ "castle": 17,
143
+ "caterpillar": 18,
144
+ "cattle": 19,
145
+ "chair": 20,
146
+ "chimpanzee": 21,
147
+ "clock": 22,
148
+ "cloud": 23,
149
+ "cockroach": 24,
150
+ "couch": 25,
151
+ "cra": 26,
152
+ "crocodile": 27,
153
+ "cup": 28,
154
+ "dinosaur": 29,
155
+ "dolphin": 30,
156
+ "elephant": 31,
157
+ "flatfish": 32,
158
+ "forest": 33,
159
+ "fox": 34,
160
+ "girl": 35,
161
+ "hamster": 36,
162
+ "house": 37,
163
+ "kangaroo": 38,
164
+ "keyboard": 39,
165
+ "lamp": 40,
166
+ "lawn_mower": 41,
167
+ "leopard": 42,
168
+ "lion": 43,
169
+ "lizard": 44,
170
+ "lobster": 45,
171
+ "man": 46,
172
+ "maple_tree": 47,
173
+ "motorcycle": 48,
174
+ "mountain": 49,
175
+ "mouse": 50,
176
+ "mushroom": 51,
177
+ "oak_tree": 52,
178
+ "orange": 53,
179
+ "orchid": 54,
180
+ "otter": 55,
181
+ "palm_tree": 56,
182
+ "pear": 57,
183
+ "pickup_truck": 58,
184
+ "pine_tree": 59,
185
+ "plain": 60,
186
+ "plate": 61,
187
+ "poppy": 62,
188
+ "porcupine": 63,
189
+ "possum": 64,
190
+ "rabbit": 65,
191
+ "raccoon": 66,
192
+ "ray": 67,
193
+ "road": 68,
194
+ "rocket": 69,
195
+ "rose": 70,
196
+ "sea": 71,
197
+ "seal": 72,
198
+ "shark": 73,
199
+ "shrew": 74,
200
+ "skunk": 75,
201
+ "skyscraper": 76,
202
+ "snail": 77,
203
+ "snake": 78,
204
+ "spider": 79,
205
+ "squirrel": 80,
206
+ "streetcar": 81,
207
+ "sunflower": 82,
208
+ "sweet_pepper": 83,
209
+ "table": 84,
210
+ "tank": 85,
211
+ "telephone": 86,
212
+ "television": 87,
213
+ "tiger": 88,
214
+ "tractor": 89,
215
+ "train": 90,
216
+ "trout": 91,
217
+ "tulip": 92,
218
+ "turtle": 93,
219
+ "wardrobe": 94,
220
+ "whale": 95,
221
+ "willow_tree": 96,
222
+ "wolf": 97,
223
+ "woman": 98,
224
+ "worm": 99
225
+ },
226
+ "layer_type": "bottleneck",
227
+ "model_type": "resnet",
228
+ "num_channels": 3,
229
+ "out_features": [
230
+ "stage4"
231
+ ],
232
+ "out_indices": [
233
+ 4
234
+ ],
235
+ "stage_names": [
236
+ "stem",
237
+ "stage1",
238
+ "stage2",
239
+ "stage3",
240
+ "stage4"
241
+ ],
242
+ "transformers_version": "4.56.1"
243
+ }
resnet152-cifar100-batch_size_64_lr_0.001_training_data_ratio_1.0-2000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_0.5-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4c5d158167a72dfc26dc5fb6d41b79e76751a7cda5a5fb22bba966365f7d394
3
+ size 234125152
resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_0.8-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe4117aa0ef17538dac431ff185479c522d99d89a8522bc16977154916abecba
3
+ size 234125152
resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_1.0-1000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset cifar100.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 64
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.005
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 1.0
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: cifar100
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ cifar100:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/cifar100
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ cifar100:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/cifar100
74
+ split: test
75
+ ```
76
+
resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_1.0-1000/config.json ADDED
@@ -0,0 +1,243 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "apple",
24
+ "1": "aquarium_fish",
25
+ "2": "baby",
26
+ "3": "bear",
27
+ "4": "beaver",
28
+ "5": "bed",
29
+ "6": "bee",
30
+ "7": "beetle",
31
+ "8": "bicycle",
32
+ "9": "bottle",
33
+ "10": "bowl",
34
+ "11": "boy",
35
+ "12": "bridge",
36
+ "13": "bus",
37
+ "14": "butterfly",
38
+ "15": "camel",
39
+ "16": "can",
40
+ "17": "castle",
41
+ "18": "caterpillar",
42
+ "19": "cattle",
43
+ "20": "chair",
44
+ "21": "chimpanzee",
45
+ "22": "clock",
46
+ "23": "cloud",
47
+ "24": "cockroach",
48
+ "25": "couch",
49
+ "26": "cra",
50
+ "27": "crocodile",
51
+ "28": "cup",
52
+ "29": "dinosaur",
53
+ "30": "dolphin",
54
+ "31": "elephant",
55
+ "32": "flatfish",
56
+ "33": "forest",
57
+ "34": "fox",
58
+ "35": "girl",
59
+ "36": "hamster",
60
+ "37": "house",
61
+ "38": "kangaroo",
62
+ "39": "keyboard",
63
+ "40": "lamp",
64
+ "41": "lawn_mower",
65
+ "42": "leopard",
66
+ "43": "lion",
67
+ "44": "lizard",
68
+ "45": "lobster",
69
+ "46": "man",
70
+ "47": "maple_tree",
71
+ "48": "motorcycle",
72
+ "49": "mountain",
73
+ "50": "mouse",
74
+ "51": "mushroom",
75
+ "52": "oak_tree",
76
+ "53": "orange",
77
+ "54": "orchid",
78
+ "55": "otter",
79
+ "56": "palm_tree",
80
+ "57": "pear",
81
+ "58": "pickup_truck",
82
+ "59": "pine_tree",
83
+ "60": "plain",
84
+ "61": "plate",
85
+ "62": "poppy",
86
+ "63": "porcupine",
87
+ "64": "possum",
88
+ "65": "rabbit",
89
+ "66": "raccoon",
90
+ "67": "ray",
91
+ "68": "road",
92
+ "69": "rocket",
93
+ "70": "rose",
94
+ "71": "sea",
95
+ "72": "seal",
96
+ "73": "shark",
97
+ "74": "shrew",
98
+ "75": "skunk",
99
+ "76": "skyscraper",
100
+ "77": "snail",
101
+ "78": "snake",
102
+ "79": "spider",
103
+ "80": "squirrel",
104
+ "81": "streetcar",
105
+ "82": "sunflower",
106
+ "83": "sweet_pepper",
107
+ "84": "table",
108
+ "85": "tank",
109
+ "86": "telephone",
110
+ "87": "television",
111
+ "88": "tiger",
112
+ "89": "tractor",
113
+ "90": "train",
114
+ "91": "trout",
115
+ "92": "tulip",
116
+ "93": "turtle",
117
+ "94": "wardrobe",
118
+ "95": "whale",
119
+ "96": "willow_tree",
120
+ "97": "wolf",
121
+ "98": "woman",
122
+ "99": "worm"
123
+ },
124
+ "label2id": {
125
+ "apple": 0,
126
+ "aquarium_fish": 1,
127
+ "baby": 2,
128
+ "bear": 3,
129
+ "beaver": 4,
130
+ "bed": 5,
131
+ "bee": 6,
132
+ "beetle": 7,
133
+ "bicycle": 8,
134
+ "bottle": 9,
135
+ "bowl": 10,
136
+ "boy": 11,
137
+ "bridge": 12,
138
+ "bus": 13,
139
+ "butterfly": 14,
140
+ "camel": 15,
141
+ "can": 16,
142
+ "castle": 17,
143
+ "caterpillar": 18,
144
+ "cattle": 19,
145
+ "chair": 20,
146
+ "chimpanzee": 21,
147
+ "clock": 22,
148
+ "cloud": 23,
149
+ "cockroach": 24,
150
+ "couch": 25,
151
+ "cra": 26,
152
+ "crocodile": 27,
153
+ "cup": 28,
154
+ "dinosaur": 29,
155
+ "dolphin": 30,
156
+ "elephant": 31,
157
+ "flatfish": 32,
158
+ "forest": 33,
159
+ "fox": 34,
160
+ "girl": 35,
161
+ "hamster": 36,
162
+ "house": 37,
163
+ "kangaroo": 38,
164
+ "keyboard": 39,
165
+ "lamp": 40,
166
+ "lawn_mower": 41,
167
+ "leopard": 42,
168
+ "lion": 43,
169
+ "lizard": 44,
170
+ "lobster": 45,
171
+ "man": 46,
172
+ "maple_tree": 47,
173
+ "motorcycle": 48,
174
+ "mountain": 49,
175
+ "mouse": 50,
176
+ "mushroom": 51,
177
+ "oak_tree": 52,
178
+ "orange": 53,
179
+ "orchid": 54,
180
+ "otter": 55,
181
+ "palm_tree": 56,
182
+ "pear": 57,
183
+ "pickup_truck": 58,
184
+ "pine_tree": 59,
185
+ "plain": 60,
186
+ "plate": 61,
187
+ "poppy": 62,
188
+ "porcupine": 63,
189
+ "possum": 64,
190
+ "rabbit": 65,
191
+ "raccoon": 66,
192
+ "ray": 67,
193
+ "road": 68,
194
+ "rocket": 69,
195
+ "rose": 70,
196
+ "sea": 71,
197
+ "seal": 72,
198
+ "shark": 73,
199
+ "shrew": 74,
200
+ "skunk": 75,
201
+ "skyscraper": 76,
202
+ "snail": 77,
203
+ "snake": 78,
204
+ "spider": 79,
205
+ "squirrel": 80,
206
+ "streetcar": 81,
207
+ "sunflower": 82,
208
+ "sweet_pepper": 83,
209
+ "table": 84,
210
+ "tank": 85,
211
+ "telephone": 86,
212
+ "television": 87,
213
+ "tiger": 88,
214
+ "tractor": 89,
215
+ "train": 90,
216
+ "trout": 91,
217
+ "tulip": 92,
218
+ "turtle": 93,
219
+ "wardrobe": 94,
220
+ "whale": 95,
221
+ "willow_tree": 96,
222
+ "wolf": 97,
223
+ "woman": 98,
224
+ "worm": 99
225
+ },
226
+ "layer_type": "bottleneck",
227
+ "model_type": "resnet",
228
+ "num_channels": 3,
229
+ "out_features": [
230
+ "stage4"
231
+ ],
232
+ "out_indices": [
233
+ 4
234
+ ],
235
+ "stage_names": [
236
+ "stem",
237
+ "stage1",
238
+ "stage2",
239
+ "stage3",
240
+ "stage4"
241
+ ],
242
+ "transformers_version": "4.56.1"
243
+ }
resnet152-cifar100-batch_size_64_lr_0.005_training_data_ratio_1.0-1000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-dtd-batch_size_128_lr_0.001_training_data_ratio_0.8-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5daeeed4429e1f632694e2cc7db3925544a943dff40b4e8364b7deddc89f5f0d
3
+ size 233690692
resnet152-dtd-batch_size_128_lr_0.005_training_data_ratio_0.8-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5beb1a19d7998a314fd85b08c5f8a0885a8c0595184d9f8ca443fc546d58b89d
3
+ size 233690692
resnet152-dtd-batch_size_128_lr_0.005_training_data_ratio_0.8-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0368b7bc8b129ef6804ecedd058e98585156a73c98bd50bc49b1a348d3c87317
3
+ size 233690692
resnet152-dtd-batch_size_128_lr_0.01_training_data_ratio_0.8-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ead068e08a451ef6c6c856faf909f8e77f9db8732a6d51fbbcd73b46f8bcc3a
3
+ size 233690692
resnet152-dtd-batch_size_256_lr_0.005_training_data_ratio_1.0-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:381947845abebb61a5139e0cdd8eefd926dc6fc1a90bd9249cc3242d3a98b1f0
3
+ size 233690692
resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_0.8-2000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset dtd.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 128
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.01
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 0.8
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: dtd
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ dtd:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/dtd
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ dtd:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/dtd
74
+ split: test
75
+ ```
76
+
resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_0.8-2000/config.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "banded",
24
+ "1": "blotchy",
25
+ "2": "braided",
26
+ "3": "bubbly",
27
+ "4": "bumpy",
28
+ "5": "chequered",
29
+ "6": "cobwebbed",
30
+ "7": "cracked",
31
+ "8": "crosshatched",
32
+ "9": "crystalline",
33
+ "10": "dotted",
34
+ "11": "fibrous",
35
+ "12": "flecked",
36
+ "13": "freckled",
37
+ "14": "frilly",
38
+ "15": "gauzy",
39
+ "16": "grid",
40
+ "17": "grooved",
41
+ "18": "honeycombed",
42
+ "19": "interlaced",
43
+ "20": "knitted",
44
+ "21": "lacelike",
45
+ "22": "lined",
46
+ "23": "marbled",
47
+ "24": "matted",
48
+ "25": "meshed",
49
+ "26": "paisley",
50
+ "27": "perforated",
51
+ "28": "pitted",
52
+ "29": "pleated",
53
+ "30": "polka-dotted",
54
+ "31": "porous",
55
+ "32": "potholed",
56
+ "33": "scaly",
57
+ "34": "smeared",
58
+ "35": "spiralled",
59
+ "36": "sprinkled",
60
+ "37": "stained",
61
+ "38": "stratified",
62
+ "39": "striped",
63
+ "40": "studded",
64
+ "41": "swirly",
65
+ "42": "veined",
66
+ "43": "waffled",
67
+ "44": "woven",
68
+ "45": "wrinkled",
69
+ "46": "zigzagged"
70
+ },
71
+ "label2id": {
72
+ "banded": 0,
73
+ "blotchy": 1,
74
+ "braided": 2,
75
+ "bubbly": 3,
76
+ "bumpy": 4,
77
+ "chequered": 5,
78
+ "cobwebbed": 6,
79
+ "cracked": 7,
80
+ "crosshatched": 8,
81
+ "crystalline": 9,
82
+ "dotted": 10,
83
+ "fibrous": 11,
84
+ "flecked": 12,
85
+ "freckled": 13,
86
+ "frilly": 14,
87
+ "gauzy": 15,
88
+ "grid": 16,
89
+ "grooved": 17,
90
+ "honeycombed": 18,
91
+ "interlaced": 19,
92
+ "knitted": 20,
93
+ "lacelike": 21,
94
+ "lined": 22,
95
+ "marbled": 23,
96
+ "matted": 24,
97
+ "meshed": 25,
98
+ "paisley": 26,
99
+ "perforated": 27,
100
+ "pitted": 28,
101
+ "pleated": 29,
102
+ "polka-dotted": 30,
103
+ "porous": 31,
104
+ "potholed": 32,
105
+ "scaly": 33,
106
+ "smeared": 34,
107
+ "spiralled": 35,
108
+ "sprinkled": 36,
109
+ "stained": 37,
110
+ "stratified": 38,
111
+ "striped": 39,
112
+ "studded": 40,
113
+ "swirly": 41,
114
+ "veined": 42,
115
+ "waffled": 43,
116
+ "woven": 44,
117
+ "wrinkled": 45,
118
+ "zigzagged": 46
119
+ },
120
+ "layer_type": "bottleneck",
121
+ "model_type": "resnet",
122
+ "num_channels": 3,
123
+ "out_features": [
124
+ "stage4"
125
+ ],
126
+ "out_indices": [
127
+ 4
128
+ ],
129
+ "stage_names": [
130
+ "stem",
131
+ "stage1",
132
+ "stage2",
133
+ "stage3",
134
+ "stage4"
135
+ ],
136
+ "transformers_version": "4.56.1"
137
+ }
resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_0.8-2000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-dtd-batch_size_256_lr_0.01_training_data_ratio_1.0-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1446ecd704aeb02706b19c59248f937375d8528abfa265055ce3394bd62dd893
3
+ size 233690692
resnet152-dtd-batch_size_64_lr_0.005_training_data_ratio_0.8-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:152682eaa37ad1660e80e12e8aecfe387c57198cebad30cae5a8b3a1a18d9394
3
+ size 233690692
resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_0.5-1000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset dtd.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 64
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.01
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: true
47
+ save_top_k: -1
48
+ training_data_ratio: 0.5
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: dtd
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ dtd:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/dtd
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ dtd:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/dtd
74
+ split: test
75
+ ```
76
+
resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_0.5-1000/config.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "banded",
24
+ "1": "blotchy",
25
+ "2": "braided",
26
+ "3": "bubbly",
27
+ "4": "bumpy",
28
+ "5": "chequered",
29
+ "6": "cobwebbed",
30
+ "7": "cracked",
31
+ "8": "crosshatched",
32
+ "9": "crystalline",
33
+ "10": "dotted",
34
+ "11": "fibrous",
35
+ "12": "flecked",
36
+ "13": "freckled",
37
+ "14": "frilly",
38
+ "15": "gauzy",
39
+ "16": "grid",
40
+ "17": "grooved",
41
+ "18": "honeycombed",
42
+ "19": "interlaced",
43
+ "20": "knitted",
44
+ "21": "lacelike",
45
+ "22": "lined",
46
+ "23": "marbled",
47
+ "24": "matted",
48
+ "25": "meshed",
49
+ "26": "paisley",
50
+ "27": "perforated",
51
+ "28": "pitted",
52
+ "29": "pleated",
53
+ "30": "polka-dotted",
54
+ "31": "porous",
55
+ "32": "potholed",
56
+ "33": "scaly",
57
+ "34": "smeared",
58
+ "35": "spiralled",
59
+ "36": "sprinkled",
60
+ "37": "stained",
61
+ "38": "stratified",
62
+ "39": "striped",
63
+ "40": "studded",
64
+ "41": "swirly",
65
+ "42": "veined",
66
+ "43": "waffled",
67
+ "44": "woven",
68
+ "45": "wrinkled",
69
+ "46": "zigzagged"
70
+ },
71
+ "label2id": {
72
+ "banded": 0,
73
+ "blotchy": 1,
74
+ "braided": 2,
75
+ "bubbly": 3,
76
+ "bumpy": 4,
77
+ "chequered": 5,
78
+ "cobwebbed": 6,
79
+ "cracked": 7,
80
+ "crosshatched": 8,
81
+ "crystalline": 9,
82
+ "dotted": 10,
83
+ "fibrous": 11,
84
+ "flecked": 12,
85
+ "freckled": 13,
86
+ "frilly": 14,
87
+ "gauzy": 15,
88
+ "grid": 16,
89
+ "grooved": 17,
90
+ "honeycombed": 18,
91
+ "interlaced": 19,
92
+ "knitted": 20,
93
+ "lacelike": 21,
94
+ "lined": 22,
95
+ "marbled": 23,
96
+ "matted": 24,
97
+ "meshed": 25,
98
+ "paisley": 26,
99
+ "perforated": 27,
100
+ "pitted": 28,
101
+ "pleated": 29,
102
+ "polka-dotted": 30,
103
+ "porous": 31,
104
+ "potholed": 32,
105
+ "scaly": 33,
106
+ "smeared": 34,
107
+ "spiralled": 35,
108
+ "sprinkled": 36,
109
+ "stained": 37,
110
+ "stratified": 38,
111
+ "striped": 39,
112
+ "studded": 40,
113
+ "swirly": 41,
114
+ "veined": 42,
115
+ "waffled": 43,
116
+ "woven": 44,
117
+ "wrinkled": 45,
118
+ "zigzagged": 46
119
+ },
120
+ "layer_type": "bottleneck",
121
+ "model_type": "resnet",
122
+ "num_channels": 3,
123
+ "out_features": [
124
+ "stage4"
125
+ ],
126
+ "out_indices": [
127
+ 4
128
+ ],
129
+ "stage_names": [
130
+ "stem",
131
+ "stage1",
132
+ "stage2",
133
+ "stage3",
134
+ "stage4"
135
+ ],
136
+ "transformers_version": "4.56.1"
137
+ }
resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_0.5-1000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_1.0-1000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset dtd.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 64
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.01
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: true
47
+ save_top_k: -1
48
+ training_data_ratio: 1.0
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: dtd
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ dtd:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/dtd
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ dtd:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/dtd
74
+ split: test
75
+ ```
76
+
resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_1.0-1000/config.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "banded",
24
+ "1": "blotchy",
25
+ "2": "braided",
26
+ "3": "bubbly",
27
+ "4": "bumpy",
28
+ "5": "chequered",
29
+ "6": "cobwebbed",
30
+ "7": "cracked",
31
+ "8": "crosshatched",
32
+ "9": "crystalline",
33
+ "10": "dotted",
34
+ "11": "fibrous",
35
+ "12": "flecked",
36
+ "13": "freckled",
37
+ "14": "frilly",
38
+ "15": "gauzy",
39
+ "16": "grid",
40
+ "17": "grooved",
41
+ "18": "honeycombed",
42
+ "19": "interlaced",
43
+ "20": "knitted",
44
+ "21": "lacelike",
45
+ "22": "lined",
46
+ "23": "marbled",
47
+ "24": "matted",
48
+ "25": "meshed",
49
+ "26": "paisley",
50
+ "27": "perforated",
51
+ "28": "pitted",
52
+ "29": "pleated",
53
+ "30": "polka-dotted",
54
+ "31": "porous",
55
+ "32": "potholed",
56
+ "33": "scaly",
57
+ "34": "smeared",
58
+ "35": "spiralled",
59
+ "36": "sprinkled",
60
+ "37": "stained",
61
+ "38": "stratified",
62
+ "39": "striped",
63
+ "40": "studded",
64
+ "41": "swirly",
65
+ "42": "veined",
66
+ "43": "waffled",
67
+ "44": "woven",
68
+ "45": "wrinkled",
69
+ "46": "zigzagged"
70
+ },
71
+ "label2id": {
72
+ "banded": 0,
73
+ "blotchy": 1,
74
+ "braided": 2,
75
+ "bubbly": 3,
76
+ "bumpy": 4,
77
+ "chequered": 5,
78
+ "cobwebbed": 6,
79
+ "cracked": 7,
80
+ "crosshatched": 8,
81
+ "crystalline": 9,
82
+ "dotted": 10,
83
+ "fibrous": 11,
84
+ "flecked": 12,
85
+ "freckled": 13,
86
+ "frilly": 14,
87
+ "gauzy": 15,
88
+ "grid": 16,
89
+ "grooved": 17,
90
+ "honeycombed": 18,
91
+ "interlaced": 19,
92
+ "knitted": 20,
93
+ "lacelike": 21,
94
+ "lined": 22,
95
+ "marbled": 23,
96
+ "matted": 24,
97
+ "meshed": 25,
98
+ "paisley": 26,
99
+ "perforated": 27,
100
+ "pitted": 28,
101
+ "pleated": 29,
102
+ "polka-dotted": 30,
103
+ "porous": 31,
104
+ "potholed": 32,
105
+ "scaly": 33,
106
+ "smeared": 34,
107
+ "spiralled": 35,
108
+ "sprinkled": 36,
109
+ "stained": 37,
110
+ "stratified": 38,
111
+ "striped": 39,
112
+ "studded": 40,
113
+ "swirly": 41,
114
+ "veined": 42,
115
+ "waffled": 43,
116
+ "woven": 44,
117
+ "wrinkled": 45,
118
+ "zigzagged": 46
119
+ },
120
+ "layer_type": "bottleneck",
121
+ "model_type": "resnet",
122
+ "num_channels": 3,
123
+ "out_features": [
124
+ "stage4"
125
+ ],
126
+ "out_indices": [
127
+ 4
128
+ ],
129
+ "stage_names": [
130
+ "stem",
131
+ "stage1",
132
+ "stage2",
133
+ "stage3",
134
+ "stage4"
135
+ ],
136
+ "transformers_version": "4.56.1"
137
+ }
resnet152-dtd-batch_size_64_lr_0.01_training_data_ratio_1.0-1000/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_pct": 0.875,
3
+ "do_normalize": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "ConvNextImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "shortest_edge": 224
21
+ }
22
+ }
resnet152-emnist_letters-batch_size_128_lr_0.001_training_data_ratio_0.8-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:905c4dd7b87cc169ff9f26cd13e8e3726983892962ca9a1028a2f036b9ae7800
3
+ size 233518560
resnet152-emnist_letters-batch_size_128_lr_0.001_training_data_ratio_1.0-4000/README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/resnet-152
3
+ library_name: transformers
4
+ tags:
5
+ - fusion-bench
6
+ - merge
7
+ ---
8
+ # Deep Model Fusion
9
+
10
+ Fine-tuned ResNet model on dataset emnist_letters.
11
+
12
+ ## Models Merged
13
+
14
+ This is a merged model created using [fusion-bench](https://github.com/tanganke/fusion_bench).
15
+
16
+ The following models were included in the merge:
17
+
18
+
19
+
20
+
21
+ ## Configuration
22
+
23
+ The following YAML configuration was used to produce this model:
24
+
25
+ ### Algorithm Configuration
26
+
27
+ ```yaml
28
+ _recursive_: false
29
+ _target_: fusion_bench.method.classification.image_classification_finetune.ImageClassificationFineTuning
30
+ _usage_: null
31
+ _version_: 0.2.25.dev0
32
+ dataloader_kwargs:
33
+ batch_size: 128
34
+ num_workers: 8
35
+ pin_memory: true
36
+ label_smoothing: 0
37
+ lr_scheduler: null
38
+ max_epochs: -1
39
+ max_steps: 4000
40
+ optimizer:
41
+ _target_: torch.optim.SGD
42
+ lr: 0.001
43
+ momentum: 0.9
44
+ weight_decay: 0.0001
45
+ save_interval: 1000
46
+ save_on_train_epoch_end: false
47
+ save_top_k: -1
48
+ training_data_ratio: 1.0
49
+ ```
50
+
51
+ ### Model Pool Configuration
52
+
53
+ ```yaml
54
+ _recursive_: false
55
+ _target_: fusion_bench.modelpool.resnet_for_image_classification.ResNetForImageClassificationPool
56
+ _usage_: null
57
+ _version_: 0.2.25.dev0
58
+ models:
59
+ _pretrained_:
60
+ config_path: microsoft/resnet-152
61
+ dataset_name: emnist_letters
62
+ pretrained: true
63
+ test_datasets: null
64
+ train_datasets:
65
+ emnist_letters:
66
+ _target_: datasets.load_dataset
67
+ path: tanganke/emnist_letters
68
+ split: train
69
+ type: transformers
70
+ val_datasets:
71
+ emnist_letters:
72
+ _target_: datasets.load_dataset
73
+ path: tanganke/emnist_letters
74
+ split: test
75
+ ```
76
+
resnet152-emnist_letters-batch_size_128_lr_0.001_training_data_ratio_1.0-4000/config.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ResNetForImageClassification"
4
+ ],
5
+ "depths": [
6
+ 3,
7
+ 8,
8
+ 36,
9
+ 3
10
+ ],
11
+ "downsample_in_bottleneck": false,
12
+ "downsample_in_first_stage": false,
13
+ "dtype": "float32",
14
+ "embedding_size": 64,
15
+ "hidden_act": "relu",
16
+ "hidden_sizes": [
17
+ 256,
18
+ 512,
19
+ 1024,
20
+ 2048
21
+ ],
22
+ "id2label": {
23
+ "0": "A",
24
+ "1": "B",
25
+ "2": "C",
26
+ "3": "D",
27
+ "4": "E",
28
+ "5": "F",
29
+ "6": "G",
30
+ "7": "H",
31
+ "8": "I",
32
+ "9": "J",
33
+ "10": "K",
34
+ "11": "L",
35
+ "12": "M",
36
+ "13": "N",
37
+ "14": "O",
38
+ "15": "P",
39
+ "16": "Q",
40
+ "17": "R",
41
+ "18": "S",
42
+ "19": "T",
43
+ "20": "U",
44
+ "21": "V",
45
+ "22": "W",
46
+ "23": "X",
47
+ "24": "Y",
48
+ "25": "Z"
49
+ },
50
+ "label2id": {
51
+ "A": 0,
52
+ "B": 1,
53
+ "C": 2,
54
+ "D": 3,
55
+ "E": 4,
56
+ "F": 5,
57
+ "G": 6,
58
+ "H": 7,
59
+ "I": 8,
60
+ "J": 9,
61
+ "K": 10,
62
+ "L": 11,
63
+ "M": 12,
64
+ "N": 13,
65
+ "O": 14,
66
+ "P": 15,
67
+ "Q": 16,
68
+ "R": 17,
69
+ "S": 18,
70
+ "T": 19,
71
+ "U": 20,
72
+ "V": 21,
73
+ "W": 22,
74
+ "X": 23,
75
+ "Y": 24,
76
+ "Z": 25
77
+ },
78
+ "layer_type": "bottleneck",
79
+ "model_type": "resnet",
80
+ "num_channels": 3,
81
+ "out_features": [
82
+ "stage4"
83
+ ],
84
+ "out_indices": [
85
+ 4
86
+ ],
87
+ "stage_names": [
88
+ "stem",
89
+ "stage1",
90
+ "stage2",
91
+ "stage3",
92
+ "stage4"
93
+ ],
94
+ "transformers_version": "4.56.1"
95
+ }