MoaazTalab commited on
Commit
4abff61
·
verified ·
1 Parent(s): c692fea

End of training

Browse files
Files changed (5) hide show
  1. README.md +26 -38
  2. config.json +15 -14
  3. model.safetensors +2 -2
  4. preprocessor_config.json +4 -9
  5. training_args.bin +1 -1
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
- license: other
4
- base_model: google/mobilenet_v2_1.0_224
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -19,17 +19,17 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # ViT_L16
21
 
22
- This model is a fine-tuned version of [google/mobilenet_v2_1.0_224](https://huggingface.co/google/mobilenet_v2_1.0_224) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.3409
25
- - Accuracy: 0.8904
26
- - Precision: 0.8746
27
- - Recall: 0.8901
28
- - F1: 0.8823
29
- - Tp: 1458
30
- - Tn: 1701
31
- - Fp: 209
32
- - Fn: 180
33
 
34
  ## Model description
35
 
@@ -49,38 +49,26 @@ More information needed
49
 
50
  The following hyperparameters were used during training:
51
  - learning_rate: 5e-06
52
- - train_batch_size: 128
53
- - eval_batch_size: 128
54
  - seed: 42
55
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
56
  - lr_scheduler_type: linear
57
- - lr_scheduler_warmup_steps: 137
58
- - num_epochs: 5
59
 
60
  ### Training results
61
 
62
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Tp | Tn | Fp | Fn |
63
- |:-------------:|:------:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:----:|:----:|:---:|:---:|
64
- | 0.6841 | 0.2432 | 27 | 0.6509 | 0.6088 | 0.5711 | 0.6129 | 0.5913 | 1004 | 1156 | 754 | 634 |
65
- | 0.6551 | 0.4865 | 54 | 0.6145 | 0.6700 | 0.6292 | 0.6941 | 0.6601 | 1137 | 1240 | 670 | 501 |
66
- | 0.5709 | 0.7297 | 81 | 0.6052 | 0.6750 | 0.6174 | 0.7784 | 0.6886 | 1275 | 1120 | 790 | 363 |
67
- | 0.4953 | 0.9730 | 108 | 0.5217 | 0.7709 | 0.6878 | 0.9225 | 0.7880 | 1511 | 1224 | 686 | 127 |
68
- | 0.4121 | 1.2162 | 135 | 0.4447 | 0.8588 | 0.8118 | 0.9035 | 0.8552 | 1480 | 1567 | 343 | 158 |
69
- | 0.3688 | 1.4595 | 162 | 0.4108 | 0.8557 | 0.8043 | 0.9084 | 0.8532 | 1488 | 1548 | 362 | 150 |
70
- | 0.3300 | 1.7027 | 189 | 0.4537 | 0.7889 | 0.7014 | 0.9451 | 0.8052 | 1548 | 1251 | 659 | 90 |
71
- | 0.3324 | 1.9459 | 216 | 0.3450 | 0.9039 | 0.8947 | 0.8974 | 0.8961 | 1470 | 1737 | 173 | 168 |
72
- | 0.3033 | 2.1892 | 243 | 0.4286 | 0.8081 | 0.7216 | 0.9512 | 0.8206 | 1558 | 1309 | 601 | 80 |
73
- | 0.3007 | 2.4324 | 270 | 0.3362 | 0.9056 | 0.9095 | 0.8834 | 0.8963 | 1447 | 1766 | 144 | 191 |
74
- | 0.2940 | 2.6757 | 297 | 0.3404 | 0.8867 | 0.8388 | 0.9341 | 0.8839 | 1530 | 1616 | 294 | 108 |
75
- | 0.2891 | 2.9189 | 324 | 0.2897 | 0.9228 | 0.9231 | 0.9084 | 0.9157 | 1488 | 1786 | 124 | 150 |
76
- | 0.2921 | 3.1622 | 351 | 0.3451 | 0.8881 | 0.9042 | 0.8474 | 0.8749 | 1388 | 1763 | 147 | 250 |
77
- | 0.2958 | 3.4054 | 378 | 0.3503 | 0.8816 | 0.8604 | 0.8877 | 0.8738 | 1454 | 1674 | 236 | 184 |
78
- | 0.2673 | 3.6486 | 405 | 0.3124 | 0.9073 | 0.8834 | 0.9206 | 0.9016 | 1508 | 1711 | 199 | 130 |
79
- | 0.2755 | 3.8919 | 432 | 0.3070 | 0.9104 | 0.9099 | 0.8944 | 0.9021 | 1465 | 1765 | 145 | 173 |
80
- | 0.2913 | 4.1351 | 459 | 0.3094 | 0.9101 | 0.8914 | 0.9170 | 0.9040 | 1502 | 1727 | 183 | 136 |
81
- | 0.2720 | 4.3784 | 486 | 0.3083 | 0.8952 | 0.9283 | 0.8376 | 0.8806 | 1372 | 1804 | 106 | 266 |
82
- | 0.2929 | 4.6216 | 513 | 0.2726 | 0.9225 | 0.9565 | 0.8718 | 0.9122 | 1428 | 1845 | 65 | 210 |
83
- | 0.2599 | 4.8649 | 540 | 0.3409 | 0.8904 | 0.8746 | 0.8901 | 0.8823 | 1458 | 1701 | 209 | 180 |
84
 
85
 
86
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
+ license: apache-2.0
4
+ base_model: google/vit-large-patch16-224
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
19
 
20
  # ViT_L16
21
 
22
+ This model is a fine-tuned version of [google/vit-large-patch16-224](https://huggingface.co/google/vit-large-patch16-224) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.0933
25
+ - Accuracy: 0.9746
26
+ - Precision: 0.9844
27
+ - Recall: 0.9603
28
+ - F1: 0.9722
29
+ - Tp: 1573
30
+ - Tn: 1885
31
+ - Fp: 25
32
+ - Fn: 65
33
 
34
  ## Model description
35
 
 
49
 
50
  The following hyperparameters were used during training:
51
  - learning_rate: 5e-06
52
+ - train_batch_size: 32
53
+ - eval_batch_size: 32
54
  - seed: 42
55
  - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
56
  - lr_scheduler_type: linear
57
+ - lr_scheduler_warmup_steps: 221
58
+ - num_epochs: 2
59
 
60
  ### Training results
61
 
62
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 | Tp | Tn | Fp | Fn |
63
+ |:-------------:|:------:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|:----:|:----:|:--:|:---:|
64
+ | 0.4292 | 0.2477 | 110 | 0.1793 | 0.9464 | 0.9925 | 0.8907 | 0.9389 | 1459 | 1899 | 11 | 179 |
65
+ | 0.2230 | 0.4955 | 220 | 0.1211 | 0.9651 | 0.9980 | 0.9261 | 0.9607 | 1517 | 1907 | 3 | 121 |
66
+ | 0.1974 | 0.7432 | 330 | 0.1342 | 0.9690 | 0.9811 | 0.9512 | 0.9659 | 1558 | 1880 | 30 | 80 |
67
+ | 0.1879 | 0.9910 | 440 | 0.1397 | 0.9628 | 0.9591 | 0.9603 | 0.9597 | 1573 | 1843 | 67 | 65 |
68
+ | 0.1643 | 1.2387 | 550 | 0.1083 | 0.9741 | 0.9856 | 0.9579 | 0.9715 | 1569 | 1887 | 23 | 69 |
69
+ | 0.1654 | 1.4865 | 660 | 0.0963 | 0.9715 | 0.9936 | 0.9444 | 0.9684 | 1547 | 1900 | 10 | 91 |
70
+ | 0.1664 | 1.7342 | 770 | 0.1130 | 0.9693 | 0.9693 | 0.9640 | 0.9666 | 1579 | 1860 | 50 | 59 |
71
+ | 0.1637 | 1.9820 | 880 | 0.0933 | 0.9746 | 0.9844 | 0.9603 | 0.9722 | 1573 | 1885 | 25 | 65 |
 
 
 
 
 
 
 
 
 
 
 
 
72
 
73
 
74
  ### Framework versions
config.json CHANGED
@@ -1,33 +1,34 @@
1
  {
2
  "architectures": [
3
- "MobileNetV2ForImageClassification"
4
  ],
5
- "classifier_dropout_prob": 0.2,
6
- "depth_divisible_by": 8,
7
- "depth_multiplier": 1.0,
8
  "dtype": "float32",
9
- "expand_ratio": 6,
10
- "finegrained_output": true,
11
- "first_layer_is_expansion": true,
12
- "hidden_act": "relu6",
13
  "id2label": {
14
  "0": "0",
15
  "1": "1"
16
  },
17
  "image_size": 224,
18
  "initializer_range": 0.02,
 
19
  "label2id": {
20
  "0": 0,
21
  "1": 1
22
  },
23
- "layer_norm_eps": 0.001,
24
- "min_depth": 8,
25
- "model_type": "mobilenet_v2",
26
  "num_channels": 3,
27
- "output_stride": 32,
 
 
 
28
  "problem_type": "single_label_classification",
29
- "semantic_loss_ignore_index": 255,
30
- "tf_padding": true,
31
  "transformers_version": "5.0.0",
32
  "use_cache": false
33
  }
 
1
  {
2
  "architectures": [
3
+ "ViTForImageClassification"
4
  ],
5
+ "attention_probs_dropout_prob": 0.0,
 
 
6
  "dtype": "float32",
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 1024,
11
  "id2label": {
12
  "0": "0",
13
  "1": "1"
14
  },
15
  "image_size": 224,
16
  "initializer_range": 0.02,
17
+ "intermediate_size": 4096,
18
  "label2id": {
19
  "0": 0,
20
  "1": 1
21
  },
22
+ "layer_norm_eps": 1e-12,
23
+ "model_type": "vit",
24
+ "num_attention_heads": 16,
25
  "num_channels": 3,
26
+ "num_hidden_layers": 24,
27
+ "patch_size": 16,
28
+ "pooler_act": "tanh",
29
+ "pooler_output_size": 1024,
30
  "problem_type": "single_label_classification",
31
+ "qkv_bias": true,
 
32
  "transformers_version": "5.0.0",
33
  "use_cache": false
34
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6e9644d68cf097846ddac5c08ca36f23f8e7ad9770058bd4d332fc127e8bce2
3
- size 9080216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:779aeafe9a53b2a1b8a4b18e8a6a8bf637f7940dbd277f1883680150cb5fee1e
3
+ size 1213261264
preprocessor_config.json CHANGED
@@ -1,12 +1,6 @@
1
  {
2
- "crop_size": {
3
- "height": 224,
4
- "width": 224
5
- },
6
- "data_format": "channels_first",
7
- "do_center_crop": true,
8
  "do_normalize": true,
9
- "do_reduce_labels": false,
10
  "do_rescale": true,
11
  "do_resize": true,
12
  "image_mean": [
@@ -14,7 +8,7 @@
14
  0.5,
15
  0.5
16
  ],
17
- "image_processor_type": "MobileNetV2ImageProcessorFast",
18
  "image_std": [
19
  0.5,
20
  0.5,
@@ -23,6 +17,7 @@
23
  "resample": 2,
24
  "rescale_factor": 0.00392156862745098,
25
  "size": {
26
- "shortest_edge": 256
 
27
  }
28
  }
 
1
  {
2
+ "do_convert_rgb": null,
 
 
 
 
 
3
  "do_normalize": true,
 
4
  "do_rescale": true,
5
  "do_resize": true,
6
  "image_mean": [
 
8
  0.5,
9
  0.5
10
  ],
11
+ "image_processor_type": "ViTImageProcessor",
12
  "image_std": [
13
  0.5,
14
  0.5,
 
17
  "resample": 2,
18
  "rescale_factor": 0.00392156862745098,
19
  "size": {
20
+ "height": 224,
21
+ "width": 224
22
  }
23
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9718d31a58ba43fb20b11cd760cd66a129ccd3519ce7e56f71efe4a0dad8c9a2
3
  size 5137
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a20650d184cd21a5ecf7b63db317856b5992087aa5fe9cbe6ebb9da5ccee247d
3
  size 5137