Shravanig commited on
Commit
fa40102
·
verified ·
1 Parent(s): 1bdccdf

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -1,75 +1,60 @@
1
- ---
2
- license: apache-2.0
3
- base_model: google/vit-base-patch16-224-in21k
4
- tags:
5
- - generated_from_trainer
6
- metrics:
7
- - precision
8
- - recall
9
- - accuracy
10
- model-index:
11
- - name: vit-fire-detection
12
- results: []
13
- ---
14
-
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
-
18
- # vit-fire-detection
19
-
20
- This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
21
- It achieves the following results on the evaluation set:
22
- - Loss: 0.0250
23
- - Precision: 0.9934
24
- - Recall: 0.9934
25
- - Accuracy: 0.9934
26
- - F1score: 0.9934
27
-
28
- ## Model description
29
-
30
- More information needed
31
-
32
- ## Intended uses & limitations
33
-
34
- More information needed
35
-
36
- ## Training and evaluation data
37
-
38
- More information needed
39
-
40
- ## Training procedure
41
-
42
- ### Training hyperparameters
43
-
44
- The following hyperparameters were used during training:
45
- - learning_rate: 0.0002
46
- - train_batch_size: 64
47
- - eval_batch_size: 64
48
- - seed: 42
49
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
- - lr_scheduler_type: linear
51
- - lr_scheduler_warmup_steps: 100
52
- - num_epochs: 10
53
-
54
- ### Training results
55
-
56
- | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | Accuracy | F1score |
57
- |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:--------:|:-------:|
58
- | 0.5906 | 1.0 | 95 | 0.0594 | 0.9855 | 0.9854 | 0.9854 | 0.9854 |
59
- | 0.0812 | 2.0 | 190 | 0.0447 | 0.9908 | 0.9907 | 0.9907 | 0.9907 |
60
- | 0.0575 | 3.0 | 285 | 0.0779 | 0.9778 | 0.9762 | 0.9762 | 0.9762 |
61
- | 0.036 | 4.0 | 380 | 0.0242 | 0.9935 | 0.9934 | 0.9934 | 0.9934 |
62
- | 0.0395 | 5.0 | 475 | 0.0283 | 0.9921 | 0.9921 | 0.9921 | 0.9921 |
63
- | 0.0225 | 6.0 | 570 | 0.0169 | 0.9947 | 0.9947 | 0.9947 | 0.9947 |
64
- | 0.0165 | 7.0 | 665 | 0.0205 | 0.9934 | 0.9934 | 0.9934 | 0.9934 |
65
- | 0.0099 | 8.0 | 760 | 0.0180 | 0.9960 | 0.9960 | 0.9960 | 0.9960 |
66
- | 0.0074 | 9.0 | 855 | 0.0255 | 0.9947 | 0.9947 | 0.9947 | 0.9947 |
67
- | 0.0055 | 10.0 | 950 | 0.0250 | 0.9934 | 0.9934 | 0.9934 | 0.9934 |
68
-
69
-
70
- ### Framework versions
71
-
72
- - Transformers 4.36.0
73
- - Pytorch 2.1.1+cpu
74
- - Datasets 2.15.0
75
- - Tokenizers 0.15.0
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/vit-base-patch16-224-in21k
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: vit-fire-detection
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # vit-fire-detection
15
+
16
+ This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - eval_loss: 0.1693
19
+ - eval_precision: 0.9500
20
+ - eval_recall: 0.9484
21
+ - eval_accuracy: 0.9484
22
+ - eval_f1score: 0.9482
23
+ - eval_runtime: 45.7316
24
+ - eval_samples_per_second: 16.531
25
+ - eval_steps_per_second: 0.525
26
+ - epoch: 5.0
27
+ - step: 950
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 0.0002
47
+ - train_batch_size: 32
48
+ - eval_batch_size: 32
49
+ - seed: 42
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 100
53
+ - num_epochs: 10
54
+
55
+ ### Framework versions
56
+
57
+ - Transformers 4.36.0
58
+ - Pytorch 2.1.1+cpu
59
+ - Datasets 2.15.0
60
+ - Tokenizers 0.15.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json CHANGED
@@ -1,34 +1,34 @@
1
- {
2
- "_name_or_path": "google/vit-base-patch16-224-in21k",
3
- "architectures": [
4
- "ViTForImageClassification"
5
- ],
6
- "attention_probs_dropout_prob": 0.0,
7
- "encoder_stride": 16,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.0,
10
- "hidden_size": 768,
11
- "id2label": {
12
- "0": "Fire",
13
- "1": "Normal",
14
- "2": "Smoke"
15
- },
16
- "image_size": 224,
17
- "initializer_range": 0.02,
18
- "intermediate_size": 3072,
19
- "label2id": {
20
- "Fire": 0,
21
- "Normal": 1,
22
- "Smoke": 2
23
- },
24
- "layer_norm_eps": 1e-12,
25
- "model_type": "vit",
26
- "num_attention_heads": 12,
27
- "num_channels": 3,
28
- "num_hidden_layers": 12,
29
- "patch_size": 32,
30
- "problem_type": "single_label_classification",
31
- "qkv_bias": true,
32
- "torch_dtype": "float32",
33
- "transformers_version": "4.36.0"
34
- }
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Fire",
13
+ "1": "Normal",
14
+ "2": "Smoke"
15
+ },
16
+ "image_size": 224,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "Fire": 0,
21
+ "Normal": 1,
22
+ "Smoke": 2
23
+ },
24
+ "layer_norm_eps": 1e-12,
25
+ "model_type": "vit",
26
+ "num_attention_heads": 8,
27
+ "num_channels": 3,
28
+ "num_hidden_layers": 12,
29
+ "patch_size": 64,
30
+ "problem_type": "single_label_classification",
31
+ "qkv_bias": true,
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.36.0"
34
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36e43ff9f1792c04e6149b703d51cd72c84e740a2adafa5e12a0a36ff448f01b
3
- size 349853372
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b63630614b17796fa0c928b0794da306033a335f459ae98d8445d23bcc33d3fe
3
+ size 378042076
preprocessor_config.json CHANGED
@@ -1,22 +1,22 @@
1
- {
2
- "do_normalize": true,
3
- "do_rescale": true,
4
- "do_resize": true,
5
- "image_mean": [
6
- 0.5,
7
- 0.5,
8
- 0.5
9
- ],
10
- "image_processor_type": "ViTFeatureExtractor",
11
- "image_std": [
12
- 0.5,
13
- 0.5,
14
- 0.5
15
- ],
16
- "resample": 2,
17
- "rescale_factor": 0.00392156862745098,
18
- "size": {
19
- "height": 224,
20
- "width": 224
21
- }
22
- }
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTFeatureExtractor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
runs/Apr04_12-15-02_Shravani/events.out.tfevents.1712213104.Shravani.24608.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:caf740efde18049cf0a7b5674eca15a6fde110f9feb561646f35d22af56fb9ae
3
- size 10732
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e030a7deb1e6147cc3cb6d8b783222bf15d2577717cc20aa0c8acf4282263391
3
+ size 15506
runs/May06_10-04-26_Shravani/events.out.tfevents.1714970067.Shravani.34392.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41627090b83c3343fe4701e805927bdbe26235aca3342c8c160d2272155a569d
3
+ size 9784
runs/May22_06-52-46_Shravani/events.out.tfevents.1716340980.Shravani.27548.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86b67f009753185f610364f8d9156a72e0de4ef833540e710357620d83aa19c2
3
+ size 5363
tmp-checkpoint-950/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Fire",
13
+ "1": "Normal",
14
+ "2": "Smoke"
15
+ },
16
+ "image_size": 224,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "Fire": 0,
21
+ "Normal": 1,
22
+ "Smoke": 2
23
+ },
24
+ "layer_norm_eps": 1e-12,
25
+ "model_type": "vit",
26
+ "num_attention_heads": 12,
27
+ "num_channels": 3,
28
+ "num_hidden_layers": 12,
29
+ "patch_size": 64,
30
+ "problem_type": "single_label_classification",
31
+ "qkv_bias": true,
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.36.0"
34
+ }
tmp-checkpoint-950/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a65b77379eca120d8e3d90989bf2e410faf837999c9266cbd24823b86c5706
3
+ size 378042076
tmp-checkpoint-950/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1edea7193a8d5e1476b5dfbf448d7aeb2d9816c98c12781d68a4e193bcf89c78
3
+ size 756199354
tmp-checkpoint-950/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTFeatureExtractor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
tmp-checkpoint-950/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:732bc5c09c2a93ecaa4de1b7485a4b0dd21972a86253092ca4e305dd3458a309
3
+ size 13990
tmp-checkpoint-950/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbbcbb5ee6fa58bcafa461d5f254dbe3cd942dd5f6dcb4603ef82163ccb9b68e
3
+ size 1064
tmp-checkpoint-950/trainer_state.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.1471855193376541,
3
+ "best_model_checkpoint": "Shravanig/vit-fire-detection\\checkpoint-570",
4
+ "epoch": 5.0,
5
+ "eval_steps": 500,
6
+ "global_step": 950,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.26,
13
+ "learning_rate": 0.0001,
14
+ "loss": 0.9828,
15
+ "step": 50
16
+ },
17
+ {
18
+ "epoch": 0.53,
19
+ "learning_rate": 0.0002,
20
+ "loss": 0.6565,
21
+ "step": 100
22
+ },
23
+ {
24
+ "epoch": 0.79,
25
+ "learning_rate": 0.00019444444444444446,
26
+ "loss": 0.4455,
27
+ "step": 150
28
+ },
29
+ {
30
+ "epoch": 1.0,
31
+ "eval_accuracy": 0.9113756613756614,
32
+ "eval_f1score": 0.909933626964354,
33
+ "eval_loss": 0.2449590563774109,
34
+ "eval_precision": 0.911608601120207,
35
+ "eval_recall": 0.9113756613756614,
36
+ "eval_runtime": 55.8003,
37
+ "eval_samples_per_second": 13.548,
38
+ "eval_steps_per_second": 0.43,
39
+ "step": 190
40
+ },
41
+ {
42
+ "epoch": 1.05,
43
+ "learning_rate": 0.00018888888888888888,
44
+ "loss": 0.3878,
45
+ "step": 200
46
+ },
47
+ {
48
+ "epoch": 1.32,
49
+ "learning_rate": 0.00018333333333333334,
50
+ "loss": 0.4043,
51
+ "step": 250
52
+ },
53
+ {
54
+ "epoch": 1.58,
55
+ "learning_rate": 0.00017777777777777779,
56
+ "loss": 0.3663,
57
+ "step": 300
58
+ },
59
+ {
60
+ "epoch": 1.84,
61
+ "learning_rate": 0.00017222222222222224,
62
+ "loss": 0.3198,
63
+ "step": 350
64
+ },
65
+ {
66
+ "epoch": 2.0,
67
+ "eval_accuracy": 0.9431216931216931,
68
+ "eval_f1score": 0.9430234733391375,
69
+ "eval_loss": 0.15777449309825897,
70
+ "eval_precision": 0.9432524221092315,
71
+ "eval_recall": 0.9431216931216931,
72
+ "eval_runtime": 58.5248,
73
+ "eval_samples_per_second": 12.918,
74
+ "eval_steps_per_second": 0.41,
75
+ "step": 380
76
+ },
77
+ {
78
+ "epoch": 2.11,
79
+ "learning_rate": 0.0001666666666666667,
80
+ "loss": 0.2777,
81
+ "step": 400
82
+ },
83
+ {
84
+ "epoch": 2.37,
85
+ "learning_rate": 0.0001611111111111111,
86
+ "loss": 0.3112,
87
+ "step": 450
88
+ },
89
+ {
90
+ "epoch": 2.63,
91
+ "learning_rate": 0.00015555555555555556,
92
+ "loss": 0.2555,
93
+ "step": 500
94
+ },
95
+ {
96
+ "epoch": 2.89,
97
+ "learning_rate": 0.00015000000000000001,
98
+ "loss": 0.2768,
99
+ "step": 550
100
+ },
101
+ {
102
+ "epoch": 3.0,
103
+ "eval_accuracy": 0.9470899470899471,
104
+ "eval_f1score": 0.9469968822663881,
105
+ "eval_loss": 0.1471855193376541,
106
+ "eval_precision": 0.9475131567523525,
107
+ "eval_recall": 0.9470899470899471,
108
+ "eval_runtime": 32.47,
109
+ "eval_samples_per_second": 23.283,
110
+ "eval_steps_per_second": 0.739,
111
+ "step": 570
112
+ },
113
+ {
114
+ "epoch": 3.16,
115
+ "learning_rate": 0.00014444444444444444,
116
+ "loss": 0.235,
117
+ "step": 600
118
+ },
119
+ {
120
+ "epoch": 3.42,
121
+ "learning_rate": 0.0001388888888888889,
122
+ "loss": 0.2715,
123
+ "step": 650
124
+ },
125
+ {
126
+ "epoch": 3.68,
127
+ "learning_rate": 0.00013333333333333334,
128
+ "loss": 0.2444,
129
+ "step": 700
130
+ },
131
+ {
132
+ "epoch": 3.95,
133
+ "learning_rate": 0.00012777777777777776,
134
+ "loss": 0.2283,
135
+ "step": 750
136
+ },
137
+ {
138
+ "epoch": 4.0,
139
+ "eval_accuracy": 0.9431216931216931,
140
+ "eval_f1score": 0.9432838391066171,
141
+ "eval_loss": 0.15458014607429504,
142
+ "eval_precision": 0.9448779078440028,
143
+ "eval_recall": 0.9431216931216931,
144
+ "eval_runtime": 59.6685,
145
+ "eval_samples_per_second": 12.67,
146
+ "eval_steps_per_second": 0.402,
147
+ "step": 760
148
+ },
149
+ {
150
+ "epoch": 4.21,
151
+ "learning_rate": 0.00012222222222222224,
152
+ "loss": 0.2014,
153
+ "step": 800
154
+ },
155
+ {
156
+ "epoch": 4.47,
157
+ "learning_rate": 0.00011666666666666668,
158
+ "loss": 0.2085,
159
+ "step": 850
160
+ },
161
+ {
162
+ "epoch": 4.74,
163
+ "learning_rate": 0.00011111111111111112,
164
+ "loss": 0.2231,
165
+ "step": 900
166
+ },
167
+ {
168
+ "epoch": 5.0,
169
+ "learning_rate": 0.00010555555555555557,
170
+ "loss": 0.201,
171
+ "step": 950
172
+ },
173
+ {
174
+ "epoch": 5.0,
175
+ "eval_accuracy": 0.9484126984126984,
176
+ "eval_f1score": 0.9482271168351205,
177
+ "eval_loss": 0.169330894947052,
178
+ "eval_precision": 0.9499665185939696,
179
+ "eval_recall": 0.9484126984126984,
180
+ "eval_runtime": 45.7316,
181
+ "eval_samples_per_second": 16.531,
182
+ "eval_steps_per_second": 0.525,
183
+ "step": 950
184
+ }
185
+ ],
186
+ "logging_steps": 50,
187
+ "max_steps": 1900,
188
+ "num_input_tokens_seen": 0,
189
+ "num_train_epochs": 10,
190
+ "save_steps": 500,
191
+ "total_flos": 2.586214904516813e+18,
192
+ "train_batch_size": 32,
193
+ "trial_name": null,
194
+ "trial_params": null
195
+ }
tmp-checkpoint-950/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4794393befeb5f69568a4ecfefe1735cfba604c435f1d5a8a1261f6b843fc7f
3
+ size 4792
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec4a79f15a63d9ba638f9b0ec493d2d92b8b6a9b0c2ff24587d8e7a60e70b477
3
  size 4792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:270b5928caae40c5bca0e7cca3827e413320f661f75cdabd732e310d4ec2bcd3
3
  size 4792