sai-prakash-c commited on
Commit
9ce87c9
·
1 Parent(s): 471f4ce

Added trained model files

Browse files
all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_f1": 0.9689867070609877,
4
+ "eval_loss": 0.10238795727491379,
5
+ "eval_runtime": 125.4849,
6
+ "eval_samples_per_second": 100.65,
7
+ "eval_steps_per_second": 1.578,
8
+ "step": 2680
9
+ }
config.json ADDED
@@ -0,0 +1,149 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/swinv2-tiny-patch4-window16-256",
3
+ "architectures": [
4
+ "Swinv2ForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "depths": [
8
+ 2,
9
+ 2,
10
+ 6,
11
+ 2
12
+ ],
13
+ "drop_path_rate": 0.1,
14
+ "embed_dim": 96,
15
+ "encoder_stride": 32,
16
+ "hidden_act": "gelu",
17
+ "hidden_dropout_prob": 0.0,
18
+ "hidden_size": 768,
19
+ "id2label": {
20
+ "0": "Speed limit (20km/h)",
21
+ "1": "Speed limit (30km/h)",
22
+ "2": "Speed limit (50km/h)",
23
+ "3": "Speed limit (60km/h)",
24
+ "4": "Speed limit (70km/h)",
25
+ "5": "Speed limit (80km/h)",
26
+ "6": "End of speed limit (80km/h)",
27
+ "7": "Speed limit (100km/h)",
28
+ "8": "Speed limit (120km/h)",
29
+ "9": "No passing",
30
+ "10": "No passing for vehicles over 3.5 metric tons",
31
+ "11": "Right-of-way at the next intersection",
32
+ "12": "Priority road",
33
+ "13": "Yield",
34
+ "14": "Stop",
35
+ "15": "No vehicles",
36
+ "16": "Vehicles over 3.5 metric tons prohibited",
37
+ "17": "No entry",
38
+ "18": "General caution",
39
+ "19": "Dangerous curve to the left",
40
+ "20": "Dangerous curve to the right",
41
+ "21": "Double curve",
42
+ "22": "Bumpy road",
43
+ "23": "Slippery road",
44
+ "24": "Road narrows on the right",
45
+ "25": "Road work",
46
+ "26": "Traffic signals",
47
+ "27": "Pedestrians",
48
+ "28": "Children crossing",
49
+ "29": "Bicycles crossing",
50
+ "30": "Beware of ice/snow",
51
+ "31": "Wild animals crossing",
52
+ "32": "End of all speed and passing limits",
53
+ "33": "Turn right ahead",
54
+ "34": "Turn left ahead",
55
+ "35": "Ahead only",
56
+ "36": "Go straight or right",
57
+ "37": "Go straight or left",
58
+ "38": "Keep right",
59
+ "39": "Keep left",
60
+ "40": "Roundabout mandatory",
61
+ "41": "End of no passing",
62
+ "42": "End of no passing by vehicles over 3.5 metric tons"
63
+ },
64
+ "image_size": 256,
65
+ "initializer_range": 0.02,
66
+ "label2id": {
67
+ "Ahead only": 35,
68
+ "Beware of ice/snow": 30,
69
+ "Bicycles crossing": 29,
70
+ "Bumpy road": 22,
71
+ "Children crossing": 28,
72
+ "Dangerous curve to the left": 19,
73
+ "Dangerous curve to the right": 20,
74
+ "Double curve": 21,
75
+ "End of all speed and passing limits": 32,
76
+ "End of no passing": 41,
77
+ "End of no passing by vehicles over 3.5 metric tons": 42,
78
+ "End of speed limit (80km/h)": 6,
79
+ "General caution": 18,
80
+ "Go straight or left": 37,
81
+ "Go straight or right": 36,
82
+ "Keep left": 39,
83
+ "Keep right": 38,
84
+ "No entry": 17,
85
+ "No passing": 9,
86
+ "No passing for vehicles over 3.5 metric tons": 10,
87
+ "No vehicles": 15,
88
+ "Pedestrians": 27,
89
+ "Priority road": 12,
90
+ "Right-of-way at the next intersection": 11,
91
+ "Road narrows on the right": 24,
92
+ "Road work": 25,
93
+ "Roundabout mandatory": 40,
94
+ "Slippery road": 23,
95
+ "Speed limit (100km/h)": 7,
96
+ "Speed limit (120km/h)": 8,
97
+ "Speed limit (20km/h)": 0,
98
+ "Speed limit (30km/h)": 1,
99
+ "Speed limit (50km/h)": 2,
100
+ "Speed limit (60km/h)": 3,
101
+ "Speed limit (70km/h)": 4,
102
+ "Speed limit (80km/h)": 5,
103
+ "Stop": 14,
104
+ "Traffic signals": 26,
105
+ "Turn left ahead": 34,
106
+ "Turn right ahead": 33,
107
+ "Vehicles over 3.5 metric tons prohibited": 16,
108
+ "Wild animals crossing": 31,
109
+ "Yield": 13
110
+ },
111
+ "layer_norm_eps": 1e-05,
112
+ "mlp_ratio": 4.0,
113
+ "model_type": "swinv2",
114
+ "num_channels": 3,
115
+ "num_heads": [
116
+ 3,
117
+ 6,
118
+ 12,
119
+ 24
120
+ ],
121
+ "num_layers": 4,
122
+ "out_features": [
123
+ "stage4"
124
+ ],
125
+ "out_indices": [
126
+ 4
127
+ ],
128
+ "patch_size": 4,
129
+ "path_norm": true,
130
+ "pretrained_window_sizes": [
131
+ 0,
132
+ 0,
133
+ 0,
134
+ 0
135
+ ],
136
+ "problem_type": "single_label_classification",
137
+ "qkv_bias": true,
138
+ "stage_names": [
139
+ "stem",
140
+ "stage1",
141
+ "stage2",
142
+ "stage3",
143
+ "stage4"
144
+ ],
145
+ "torch_dtype": "float32",
146
+ "transformers_version": "4.41.2",
147
+ "use_absolute_embeddings": false,
148
+ "window_size": 16
149
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c47afd4244906a668dda1b832ca8698b7626ff21ddf0a8e9825f6d27bab4e9f
3
+ size 110476284
trained_model_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_f1": 0.9689867070609877,
4
+ "eval_loss": 0.10238795727491379,
5
+ "eval_runtime": 125.4849,
6
+ "eval_samples_per_second": 100.65,
7
+ "eval_steps_per_second": 1.578,
8
+ "step": 2680
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,371 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.06405104696750641,
3
+ "best_model_checkpoint": "./trained_model/checkpoint-2144",
4
+ "epoch": 20.0,
5
+ "eval_steps": 500,
6
+ "global_step": 2680,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 35.17252731323242,
14
+ "learning_rate": 5e-06,
15
+ "loss": 3.7473,
16
+ "step": 134
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_f1": 0.04727812496565904,
21
+ "eval_loss": 3.678004264831543,
22
+ "eval_runtime": 4.792,
23
+ "eval_samples_per_second": 98.706,
24
+ "eval_steps_per_second": 1.669,
25
+ "step": 134
26
+ },
27
+ {
28
+ "epoch": 2.0,
29
+ "grad_norm": 199.6754913330078,
30
+ "learning_rate": 1e-05,
31
+ "loss": 3.2514,
32
+ "step": 268
33
+ },
34
+ {
35
+ "epoch": 2.0,
36
+ "eval_f1": 0.266607271215907,
37
+ "eval_loss": 2.5477890968322754,
38
+ "eval_runtime": 4.7749,
39
+ "eval_samples_per_second": 99.059,
40
+ "eval_steps_per_second": 1.675,
41
+ "step": 268
42
+ },
43
+ {
44
+ "epoch": 3.0,
45
+ "grad_norm": 171.65182495117188,
46
+ "learning_rate": 9.444444444444445e-06,
47
+ "loss": 1.8984,
48
+ "step": 402
49
+ },
50
+ {
51
+ "epoch": 3.0,
52
+ "eval_f1": 0.7324588125106856,
53
+ "eval_loss": 1.1185234785079956,
54
+ "eval_runtime": 4.6545,
55
+ "eval_samples_per_second": 101.621,
56
+ "eval_steps_per_second": 1.719,
57
+ "step": 402
58
+ },
59
+ {
60
+ "epoch": 4.0,
61
+ "grad_norm": 35.59435272216797,
62
+ "learning_rate": 8.888888888888888e-06,
63
+ "loss": 1.0211,
64
+ "step": 536
65
+ },
66
+ {
67
+ "epoch": 4.0,
68
+ "eval_f1": 0.860704216321126,
69
+ "eval_loss": 0.5424026846885681,
70
+ "eval_runtime": 4.7373,
71
+ "eval_samples_per_second": 99.846,
72
+ "eval_steps_per_second": 1.689,
73
+ "step": 536
74
+ },
75
+ {
76
+ "epoch": 5.0,
77
+ "grad_norm": 3.402517318725586,
78
+ "learning_rate": 8.333333333333334e-06,
79
+ "loss": 0.6218,
80
+ "step": 670
81
+ },
82
+ {
83
+ "epoch": 5.0,
84
+ "eval_f1": 0.9176158056018094,
85
+ "eval_loss": 0.29218313097953796,
86
+ "eval_runtime": 4.6529,
87
+ "eval_samples_per_second": 101.658,
88
+ "eval_steps_per_second": 1.719,
89
+ "step": 670
90
+ },
91
+ {
92
+ "epoch": 6.0,
93
+ "grad_norm": 2.050248384475708,
94
+ "learning_rate": 7.77777777777778e-06,
95
+ "loss": 0.4323,
96
+ "step": 804
97
+ },
98
+ {
99
+ "epoch": 6.0,
100
+ "eval_f1": 0.9496526464988996,
101
+ "eval_loss": 0.19649288058280945,
102
+ "eval_runtime": 4.695,
103
+ "eval_samples_per_second": 100.745,
104
+ "eval_steps_per_second": 1.704,
105
+ "step": 804
106
+ },
107
+ {
108
+ "epoch": 7.0,
109
+ "grad_norm": 0.15155179798603058,
110
+ "learning_rate": 7.222222222222223e-06,
111
+ "loss": 0.3037,
112
+ "step": 938
113
+ },
114
+ {
115
+ "epoch": 7.0,
116
+ "eval_f1": 0.9658213272069875,
117
+ "eval_loss": 0.1448889821767807,
118
+ "eval_runtime": 4.6742,
119
+ "eval_samples_per_second": 101.194,
120
+ "eval_steps_per_second": 1.712,
121
+ "step": 938
122
+ },
123
+ {
124
+ "epoch": 8.0,
125
+ "grad_norm": 1.328462839126587,
126
+ "learning_rate": 6.666666666666667e-06,
127
+ "loss": 0.2473,
128
+ "step": 1072
129
+ },
130
+ {
131
+ "epoch": 8.0,
132
+ "eval_f1": 0.9722665265768234,
133
+ "eval_loss": 0.11456680297851562,
134
+ "eval_runtime": 4.6339,
135
+ "eval_samples_per_second": 102.073,
136
+ "eval_steps_per_second": 1.726,
137
+ "step": 1072
138
+ },
139
+ {
140
+ "epoch": 9.0,
141
+ "grad_norm": 313.9409484863281,
142
+ "learning_rate": 6.111111111111112e-06,
143
+ "loss": 0.2064,
144
+ "step": 1206
145
+ },
146
+ {
147
+ "epoch": 9.0,
148
+ "eval_f1": 0.9780034572033636,
149
+ "eval_loss": 0.09856382757425308,
150
+ "eval_runtime": 4.6934,
151
+ "eval_samples_per_second": 100.779,
152
+ "eval_steps_per_second": 1.705,
153
+ "step": 1206
154
+ },
155
+ {
156
+ "epoch": 10.0,
157
+ "grad_norm": 189.65391540527344,
158
+ "learning_rate": 5.555555555555557e-06,
159
+ "loss": 0.1533,
160
+ "step": 1340
161
+ },
162
+ {
163
+ "epoch": 10.0,
164
+ "eval_f1": 0.9778725803283007,
165
+ "eval_loss": 0.08937614411115646,
166
+ "eval_runtime": 4.6566,
167
+ "eval_samples_per_second": 101.577,
168
+ "eval_steps_per_second": 1.718,
169
+ "step": 1340
170
+ },
171
+ {
172
+ "epoch": 11.0,
173
+ "grad_norm": 22.184070587158203,
174
+ "learning_rate": 5e-06,
175
+ "loss": 0.1395,
176
+ "step": 1474
177
+ },
178
+ {
179
+ "epoch": 11.0,
180
+ "eval_f1": 0.9807261483480538,
181
+ "eval_loss": 0.0832449197769165,
182
+ "eval_runtime": 4.6769,
183
+ "eval_samples_per_second": 101.135,
184
+ "eval_steps_per_second": 1.711,
185
+ "step": 1474
186
+ },
187
+ {
188
+ "epoch": 12.0,
189
+ "grad_norm": 7.0296196937561035,
190
+ "learning_rate": 4.444444444444444e-06,
191
+ "loss": 0.1249,
192
+ "step": 1608
193
+ },
194
+ {
195
+ "epoch": 12.0,
196
+ "eval_f1": 0.9865706513457244,
197
+ "eval_loss": 0.07267424464225769,
198
+ "eval_runtime": 4.6735,
199
+ "eval_samples_per_second": 101.21,
200
+ "eval_steps_per_second": 1.712,
201
+ "step": 1608
202
+ },
203
+ {
204
+ "epoch": 13.0,
205
+ "grad_norm": 0.09863970428705215,
206
+ "learning_rate": 3.88888888888889e-06,
207
+ "loss": 0.1238,
208
+ "step": 1742
209
+ },
210
+ {
211
+ "epoch": 13.0,
212
+ "eval_f1": 0.9886923352646306,
213
+ "eval_loss": 0.0811920091509819,
214
+ "eval_runtime": 4.7013,
215
+ "eval_samples_per_second": 100.611,
216
+ "eval_steps_per_second": 1.702,
217
+ "step": 1742
218
+ },
219
+ {
220
+ "epoch": 14.0,
221
+ "grad_norm": 0.17580562829971313,
222
+ "learning_rate": 3.3333333333333333e-06,
223
+ "loss": 0.1103,
224
+ "step": 1876
225
+ },
226
+ {
227
+ "epoch": 14.0,
228
+ "eval_f1": 0.9887811136168541,
229
+ "eval_loss": 0.07010301947593689,
230
+ "eval_runtime": 4.6405,
231
+ "eval_samples_per_second": 101.93,
232
+ "eval_steps_per_second": 1.724,
233
+ "step": 1876
234
+ },
235
+ {
236
+ "epoch": 15.0,
237
+ "grad_norm": 1.5174826383590698,
238
+ "learning_rate": 2.7777777777777783e-06,
239
+ "loss": 0.0953,
240
+ "step": 2010
241
+ },
242
+ {
243
+ "epoch": 15.0,
244
+ "eval_f1": 0.9882510714507506,
245
+ "eval_loss": 0.07096822559833527,
246
+ "eval_runtime": 4.6554,
247
+ "eval_samples_per_second": 101.602,
248
+ "eval_steps_per_second": 1.718,
249
+ "step": 2010
250
+ },
251
+ {
252
+ "epoch": 16.0,
253
+ "grad_norm": 0.018986046314239502,
254
+ "learning_rate": 2.222222222222222e-06,
255
+ "loss": 0.0831,
256
+ "step": 2144
257
+ },
258
+ {
259
+ "epoch": 16.0,
260
+ "eval_f1": 0.9904615337218803,
261
+ "eval_loss": 0.06405104696750641,
262
+ "eval_runtime": 4.7164,
263
+ "eval_samples_per_second": 100.288,
264
+ "eval_steps_per_second": 1.696,
265
+ "step": 2144
266
+ },
267
+ {
268
+ "epoch": 17.0,
269
+ "grad_norm": 0.7639440298080444,
270
+ "learning_rate": 1.6666666666666667e-06,
271
+ "loss": 0.0813,
272
+ "step": 2278
273
+ },
274
+ {
275
+ "epoch": 17.0,
276
+ "eval_f1": 0.9904615337218803,
277
+ "eval_loss": 0.06663929671049118,
278
+ "eval_runtime": 4.6329,
279
+ "eval_samples_per_second": 102.095,
280
+ "eval_steps_per_second": 1.727,
281
+ "step": 2278
282
+ },
283
+ {
284
+ "epoch": 18.0,
285
+ "grad_norm": 0.013097506947815418,
286
+ "learning_rate": 1.111111111111111e-06,
287
+ "loss": 0.085,
288
+ "step": 2412
289
+ },
290
+ {
291
+ "epoch": 18.0,
292
+ "eval_f1": 0.9904615337218803,
293
+ "eval_loss": 0.06558458507061005,
294
+ "eval_runtime": 4.6458,
295
+ "eval_samples_per_second": 101.813,
296
+ "eval_steps_per_second": 1.722,
297
+ "step": 2412
298
+ },
299
+ {
300
+ "epoch": 19.0,
301
+ "grad_norm": 0.7450041770935059,
302
+ "learning_rate": 5.555555555555555e-07,
303
+ "loss": 0.0716,
304
+ "step": 2546
305
+ },
306
+ {
307
+ "epoch": 19.0,
308
+ "eval_f1": 0.9904615337218803,
309
+ "eval_loss": 0.06490221619606018,
310
+ "eval_runtime": 4.6882,
311
+ "eval_samples_per_second": 100.891,
312
+ "eval_steps_per_second": 1.706,
313
+ "step": 2546
314
+ },
315
+ {
316
+ "epoch": 20.0,
317
+ "grad_norm": 0.07539009302854538,
318
+ "learning_rate": 0.0,
319
+ "loss": 0.084,
320
+ "step": 2680
321
+ },
322
+ {
323
+ "epoch": 20.0,
324
+ "eval_f1": 0.9904615337218803,
325
+ "eval_loss": 0.06570780277252197,
326
+ "eval_runtime": 4.6524,
327
+ "eval_samples_per_second": 101.669,
328
+ "eval_steps_per_second": 1.72,
329
+ "step": 2680
330
+ },
331
+ {
332
+ "epoch": 20.0,
333
+ "step": 2680,
334
+ "total_flos": 2.773139351595909e+18,
335
+ "train_loss": 0.6440982267038146,
336
+ "train_runtime": 2307.5728,
337
+ "train_samples_per_second": 36.896,
338
+ "train_steps_per_second": 1.161
339
+ },
340
+ {
341
+ "epoch": 20.0,
342
+ "eval_f1": 0.9689867070609877,
343
+ "eval_loss": 0.10238795727491379,
344
+ "eval_runtime": 125.4849,
345
+ "eval_samples_per_second": 100.65,
346
+ "eval_steps_per_second": 1.578,
347
+ "step": 2680
348
+ }
349
+ ],
350
+ "logging_steps": 500,
351
+ "max_steps": 2680,
352
+ "num_input_tokens_seen": 0,
353
+ "num_train_epochs": 20,
354
+ "save_steps": 500,
355
+ "stateful_callbacks": {
356
+ "TrainerControl": {
357
+ "args": {
358
+ "should_epoch_stop": false,
359
+ "should_evaluate": false,
360
+ "should_log": false,
361
+ "should_save": true,
362
+ "should_training_stop": true
363
+ },
364
+ "attributes": {}
365
+ }
366
+ },
367
+ "total_flos": 2.773139351595909e+18,
368
+ "train_batch_size": 32,
369
+ "trial_name": null,
370
+ "trial_params": null
371
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4c07dd95c08cd618e082874c9a26aeef3ac6a0a356838847e4fd19e123f2e94
3
+ size 5048