djbp commited on
Commit
4df4aff
·
verified ·
1 Parent(s): 6c7984c

End of training

Browse files
README.md CHANGED
@@ -22,7 +22,7 @@ model-index:
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
- value: 0.8807785888077859
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +32,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.2599
36
- - Accuracy: 0.8808
37
 
38
  ## Model description
39
 
 
22
  metrics:
23
  - name: Accuracy
24
  type: accuracy
25
+ value: 0.8856447688564477
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
32
 
33
  This model is a fine-tuned version of [microsoft/swin-tiny-patch4-window7-224](https://huggingface.co/microsoft/swin-tiny-patch4-window7-224) on the imagefolder dataset.
34
  It achieves the following results on the evaluation set:
35
+ - Loss: 0.2637
36
+ - Accuracy: 0.8856
37
 
38
  ## Model description
39
 
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
- "epoch": 8.88888888888889,
3
- "eval_accuracy": 0.3924731182795699,
4
- "eval_loss": 2.05673885345459,
5
- "eval_runtime": 21.7695,
6
- "eval_samples_per_second": 8.544,
7
- "eval_steps_per_second": 0.138,
8
- "total_flos": 3.7044103653339955e+17,
9
- "train_loss": 2.3554419835408527,
10
- "train_runtime": 1715.7637,
11
- "train_samples_per_second": 9.733,
12
- "train_steps_per_second": 0.035
13
  }
 
1
  {
2
+ "epoch": 9.836065573770492,
3
+ "eval_accuracy": 0.8856447688564477,
4
+ "eval_loss": 0.26367661356925964,
5
+ "eval_runtime": 46.9001,
6
+ "eval_samples_per_second": 8.763,
7
+ "eval_steps_per_second": 0.149,
8
+ "total_flos": 1.9063087726729052e+18,
9
+ "train_loss": 0.3090126927693685,
10
+ "train_runtime": 8089.4797,
11
+ "train_samples_per_second": 9.637,
12
+ "train_steps_per_second": 0.037
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 8.88888888888889,
3
- "eval_accuracy": 0.3924731182795699,
4
- "eval_loss": 2.05673885345459,
5
- "eval_runtime": 21.7695,
6
- "eval_samples_per_second": 8.544,
7
- "eval_steps_per_second": 0.138
8
  }
 
1
  {
2
+ "epoch": 9.836065573770492,
3
+ "eval_accuracy": 0.8856447688564477,
4
+ "eval_loss": 0.26367661356925964,
5
+ "eval_runtime": 46.9001,
6
+ "eval_samples_per_second": 8.763,
7
+ "eval_steps_per_second": 0.149
8
  }
runs/Jul04_10-57-49_fc6b70f312e9/events.out.tfevents.1720098814.fc6b70f312e9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e746af3c883d2c7a3c6041506e60c5fce647e40f40c0fb5100a56df368cb103
3
+ size 140
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 8.88888888888889,
3
- "total_flos": 3.7044103653339955e+17,
4
- "train_loss": 2.3554419835408527,
5
- "train_runtime": 1715.7637,
6
- "train_samples_per_second": 9.733,
7
- "train_steps_per_second": 0.035
8
  }
 
1
  {
2
+ "epoch": 9.836065573770492,
3
+ "total_flos": 1.9063087726729052e+18,
4
+ "train_loss": 0.3090126927693685,
5
+ "train_runtime": 8089.4797,
6
+ "train_samples_per_second": 9.637,
7
+ "train_steps_per_second": 0.037
8
  }
trainer_state.json CHANGED
@@ -1,148 +1,325 @@
1
  {
2
- "best_metric": 0.3924731182795699,
3
- "best_model_checkpoint": "swin-tiny-patch4-window7-224-Mid-NonMidMarket-Classification/checkpoint-47",
4
- "epoch": 8.88888888888889,
5
  "eval_steps": 500,
6
- "global_step": 60,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.8888888888888888,
13
- "eval_accuracy": 0.13978494623655913,
14
- "eval_loss": 2.8975272178649902,
15
- "eval_runtime": 21.8981,
16
- "eval_samples_per_second": 8.494,
17
- "eval_steps_per_second": 0.137,
18
- "step": 6
19
  },
20
  {
21
- "epoch": 1.4814814814814814,
22
- "grad_norm": 3.0052640438079834,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "learning_rate": 4.62962962962963e-05,
24
- "loss": 2.9658,
25
- "step": 10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  },
27
  {
28
- "epoch": 1.925925925925926,
29
- "eval_accuracy": 0.22043010752688172,
30
- "eval_loss": 2.6865546703338623,
31
- "eval_runtime": 21.7609,
32
- "eval_samples_per_second": 8.547,
33
- "eval_steps_per_second": 0.138,
34
- "step": 13
35
  },
36
  {
37
- "epoch": 2.962962962962963,
38
- "grad_norm": 4.745502948760986,
39
  "learning_rate": 3.7037037037037037e-05,
40
- "loss": 2.6529,
41
- "step": 20
42
  },
43
  {
44
- "epoch": 2.962962962962963,
45
- "eval_accuracy": 0.3010752688172043,
46
- "eval_loss": 2.4369595050811768,
47
- "eval_runtime": 21.5886,
48
- "eval_samples_per_second": 8.616,
49
- "eval_steps_per_second": 0.139,
50
- "step": 20
 
 
 
 
 
51
  },
52
  {
53
  "epoch": 4.0,
54
- "eval_accuracy": 0.34946236559139787,
55
- "eval_loss": 2.2516093254089355,
56
- "eval_runtime": 21.6045,
57
- "eval_samples_per_second": 8.609,
58
- "eval_steps_per_second": 0.139,
59
- "step": 27
 
 
 
 
 
 
 
 
 
 
 
 
 
 
60
  },
61
  {
62
- "epoch": 4.444444444444445,
63
- "grad_norm": 6.134488105773926,
64
  "learning_rate": 2.777777777777778e-05,
65
- "loss": 2.3311,
66
- "step": 30
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67
  },
68
  {
69
- "epoch": 4.888888888888889,
70
- "eval_accuracy": 0.3709677419354839,
71
- "eval_loss": 2.1685309410095215,
72
- "eval_runtime": 21.6954,
73
- "eval_samples_per_second": 8.573,
74
- "eval_steps_per_second": 0.138,
75
- "step": 33
76
  },
77
  {
78
- "epoch": 5.925925925925926,
79
- "grad_norm": 7.691503524780273,
 
 
 
 
 
 
 
80
  "learning_rate": 1.8518518518518518e-05,
81
- "loss": 2.1441,
82
- "step": 40
83
  },
84
  {
85
- "epoch": 5.925925925925926,
86
- "eval_accuracy": 0.3655913978494624,
87
- "eval_loss": 2.0986852645874023,
88
- "eval_runtime": 21.7432,
89
- "eval_samples_per_second": 8.554,
90
- "eval_steps_per_second": 0.138,
91
- "step": 40
92
  },
93
  {
94
- "epoch": 6.962962962962963,
95
- "eval_accuracy": 0.3924731182795699,
96
- "eval_loss": 2.05673885345459,
97
- "eval_runtime": 21.4068,
98
- "eval_samples_per_second": 8.689,
99
- "eval_steps_per_second": 0.14,
100
- "step": 47
101
  },
102
  {
103
- "epoch": 7.407407407407407,
104
- "grad_norm": 6.981929302215576,
105
- "learning_rate": 9.259259259259259e-06,
106
- "loss": 2.0507,
107
- "step": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
108
  },
109
  {
110
  "epoch": 8.0,
111
- "eval_accuracy": 0.3870967741935484,
112
- "eval_loss": 2.041614055633545,
113
- "eval_runtime": 21.6481,
114
- "eval_samples_per_second": 8.592,
115
- "eval_steps_per_second": 0.139,
116
- "step": 54
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
117
  },
118
  {
119
- "epoch": 8.88888888888889,
120
- "grad_norm": 5.651075839996338,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
121
  "learning_rate": 0.0,
122
- "loss": 1.988,
123
- "step": 60
124
  },
125
  {
126
- "epoch": 8.88888888888889,
127
- "eval_accuracy": 0.3763440860215054,
128
- "eval_loss": 2.0368075370788574,
129
- "eval_runtime": 21.6178,
130
- "eval_samples_per_second": 8.604,
131
- "eval_steps_per_second": 0.139,
132
- "step": 60
133
  },
134
  {
135
- "epoch": 8.88888888888889,
136
- "step": 60,
137
- "total_flos": 3.7044103653339955e+17,
138
- "train_loss": 2.3554419835408527,
139
- "train_runtime": 1715.7637,
140
- "train_samples_per_second": 9.733,
141
- "train_steps_per_second": 0.035
142
  }
143
  ],
144
  "logging_steps": 10,
145
- "max_steps": 60,
146
  "num_input_tokens_seen": 0,
147
  "num_train_epochs": 10,
148
  "save_steps": 500,
@@ -158,7 +335,7 @@
158
  "attributes": {}
159
  }
160
  },
161
- "total_flos": 3.7044103653339955e+17,
162
  "train_batch_size": 64,
163
  "trial_name": null,
164
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.8856447688564477,
3
+ "best_model_checkpoint": "swin-tiny-patch4-window7-224-Mid-NonMidMarket-Classification/checkpoint-183",
4
+ "epoch": 9.836065573770492,
5
  "eval_steps": 500,
6
+ "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.32786885245901637,
13
+ "grad_norm": 4.936834812164307,
14
+ "learning_rate": 1.6666666666666667e-05,
15
+ "loss": 0.9945,
16
+ "step": 10
 
 
17
  },
18
  {
19
+ "epoch": 0.6557377049180327,
20
+ "grad_norm": 2.4745380878448486,
21
+ "learning_rate": 3.3333333333333335e-05,
22
+ "loss": 0.5619,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.9836065573770492,
27
+ "grad_norm": 3.340874433517456,
28
+ "learning_rate": 5e-05,
29
+ "loss": 0.4494,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.9836065573770492,
34
+ "eval_accuracy": 0.8223844282238443,
35
+ "eval_loss": 0.495254784822464,
36
+ "eval_runtime": 47.0869,
37
+ "eval_samples_per_second": 8.729,
38
+ "eval_steps_per_second": 0.149,
39
+ "step": 30
40
+ },
41
+ {
42
+ "epoch": 1.3114754098360657,
43
+ "grad_norm": 2.102787733078003,
44
+ "learning_rate": 4.814814814814815e-05,
45
+ "loss": 0.3852,
46
+ "step": 40
47
+ },
48
+ {
49
+ "epoch": 1.639344262295082,
50
+ "grad_norm": 1.9069348573684692,
51
  "learning_rate": 4.62962962962963e-05,
52
+ "loss": 0.3536,
53
+ "step": 50
54
+ },
55
+ {
56
+ "epoch": 1.9672131147540983,
57
+ "grad_norm": 2.154897451400757,
58
+ "learning_rate": 4.4444444444444447e-05,
59
+ "loss": 0.3369,
60
+ "step": 60
61
+ },
62
+ {
63
+ "epoch": 2.0,
64
+ "eval_accuracy": 0.8832116788321168,
65
+ "eval_loss": 0.3206270635128021,
66
+ "eval_runtime": 46.9296,
67
+ "eval_samples_per_second": 8.758,
68
+ "eval_steps_per_second": 0.149,
69
+ "step": 61
70
+ },
71
+ {
72
+ "epoch": 2.2950819672131146,
73
+ "grad_norm": 2.829212188720703,
74
+ "learning_rate": 4.259259259259259e-05,
75
+ "loss": 0.3171,
76
+ "step": 70
77
+ },
78
+ {
79
+ "epoch": 2.6229508196721314,
80
+ "grad_norm": 2.4426212310791016,
81
+ "learning_rate": 4.074074074074074e-05,
82
+ "loss": 0.3019,
83
+ "step": 80
84
+ },
85
+ {
86
+ "epoch": 2.9508196721311473,
87
+ "grad_norm": 2.601443290710449,
88
+ "learning_rate": 3.888888888888889e-05,
89
+ "loss": 0.3173,
90
+ "step": 90
91
  },
92
  {
93
+ "epoch": 2.9836065573770494,
94
+ "eval_accuracy": 0.8734793187347932,
95
+ "eval_loss": 0.3022870123386383,
96
+ "eval_runtime": 47.009,
97
+ "eval_samples_per_second": 8.743,
98
+ "eval_steps_per_second": 0.149,
99
+ "step": 91
100
  },
101
  {
102
+ "epoch": 3.278688524590164,
103
+ "grad_norm": 1.8936508893966675,
104
  "learning_rate": 3.7037037037037037e-05,
105
+ "loss": 0.3082,
106
+ "step": 100
107
  },
108
  {
109
+ "epoch": 3.6065573770491803,
110
+ "grad_norm": 2.6445701122283936,
111
+ "learning_rate": 3.518518518518519e-05,
112
+ "loss": 0.2837,
113
+ "step": 110
114
+ },
115
+ {
116
+ "epoch": 3.9344262295081966,
117
+ "grad_norm": 4.797309875488281,
118
+ "learning_rate": 3.3333333333333335e-05,
119
+ "loss": 0.2865,
120
+ "step": 120
121
  },
122
  {
123
  "epoch": 4.0,
124
+ "eval_accuracy": 0.8734793187347932,
125
+ "eval_loss": 0.2900189161300659,
126
+ "eval_runtime": 46.7107,
127
+ "eval_samples_per_second": 8.799,
128
+ "eval_steps_per_second": 0.15,
129
+ "step": 122
130
+ },
131
+ {
132
+ "epoch": 4.262295081967213,
133
+ "grad_norm": 2.120328664779663,
134
+ "learning_rate": 3.148148148148148e-05,
135
+ "loss": 0.2838,
136
+ "step": 130
137
+ },
138
+ {
139
+ "epoch": 4.590163934426229,
140
+ "grad_norm": 2.334688663482666,
141
+ "learning_rate": 2.962962962962963e-05,
142
+ "loss": 0.2709,
143
+ "step": 140
144
  },
145
  {
146
+ "epoch": 4.918032786885246,
147
+ "grad_norm": 2.006906747817993,
148
  "learning_rate": 2.777777777777778e-05,
149
+ "loss": 0.2612,
150
+ "step": 150
151
+ },
152
+ {
153
+ "epoch": 4.983606557377049,
154
+ "eval_accuracy": 0.8710462287104623,
155
+ "eval_loss": 0.28622761368751526,
156
+ "eval_runtime": 47.1214,
157
+ "eval_samples_per_second": 8.722,
158
+ "eval_steps_per_second": 0.149,
159
+ "step": 152
160
+ },
161
+ {
162
+ "epoch": 5.245901639344262,
163
+ "grad_norm": 2.6484014987945557,
164
+ "learning_rate": 2.5925925925925925e-05,
165
+ "loss": 0.2525,
166
+ "step": 160
167
+ },
168
+ {
169
+ "epoch": 5.573770491803279,
170
+ "grad_norm": 2.379870891571045,
171
+ "learning_rate": 2.4074074074074074e-05,
172
+ "loss": 0.2548,
173
+ "step": 170
174
+ },
175
+ {
176
+ "epoch": 5.901639344262295,
177
+ "grad_norm": 2.97287917137146,
178
+ "learning_rate": 2.2222222222222223e-05,
179
+ "loss": 0.2412,
180
+ "step": 180
181
  },
182
  {
183
+ "epoch": 6.0,
184
+ "eval_accuracy": 0.8856447688564477,
185
+ "eval_loss": 0.26367661356925964,
186
+ "eval_runtime": 46.9018,
187
+ "eval_samples_per_second": 8.763,
188
+ "eval_steps_per_second": 0.149,
189
+ "step": 183
190
  },
191
  {
192
+ "epoch": 6.229508196721311,
193
+ "grad_norm": 2.9260308742523193,
194
+ "learning_rate": 2.037037037037037e-05,
195
+ "loss": 0.246,
196
+ "step": 190
197
+ },
198
+ {
199
+ "epoch": 6.557377049180328,
200
+ "grad_norm": 3.2257614135742188,
201
  "learning_rate": 1.8518518518518518e-05,
202
+ "loss": 0.2434,
203
+ "step": 200
204
  },
205
  {
206
+ "epoch": 6.885245901639344,
207
+ "grad_norm": 2.4753129482269287,
208
+ "learning_rate": 1.6666666666666667e-05,
209
+ "loss": 0.2533,
210
+ "step": 210
 
 
211
  },
212
  {
213
+ "epoch": 6.983606557377049,
214
+ "eval_accuracy": 0.878345498783455,
215
+ "eval_loss": 0.27688971161842346,
216
+ "eval_runtime": 46.8021,
217
+ "eval_samples_per_second": 8.782,
218
+ "eval_steps_per_second": 0.15,
219
+ "step": 213
220
  },
221
  {
222
+ "epoch": 7.213114754098361,
223
+ "grad_norm": 2.772411584854126,
224
+ "learning_rate": 1.4814814814814815e-05,
225
+ "loss": 0.2399,
226
+ "step": 220
227
+ },
228
+ {
229
+ "epoch": 7.540983606557377,
230
+ "grad_norm": 2.4606740474700928,
231
+ "learning_rate": 1.2962962962962962e-05,
232
+ "loss": 0.2454,
233
+ "step": 230
234
+ },
235
+ {
236
+ "epoch": 7.868852459016393,
237
+ "grad_norm": 2.5486361980438232,
238
+ "learning_rate": 1.1111111111111112e-05,
239
+ "loss": 0.2413,
240
+ "step": 240
241
  },
242
  {
243
  "epoch": 8.0,
244
+ "eval_accuracy": 0.8832116788321168,
245
+ "eval_loss": 0.26485782861709595,
246
+ "eval_runtime": 47.0631,
247
+ "eval_samples_per_second": 8.733,
248
+ "eval_steps_per_second": 0.149,
249
+ "step": 244
250
+ },
251
+ {
252
+ "epoch": 8.19672131147541,
253
+ "grad_norm": 4.631506443023682,
254
+ "learning_rate": 9.259259259259259e-06,
255
+ "loss": 0.2199,
256
+ "step": 250
257
+ },
258
+ {
259
+ "epoch": 8.524590163934427,
260
+ "grad_norm": 2.3710556030273438,
261
+ "learning_rate": 7.4074074074074075e-06,
262
+ "loss": 0.2372,
263
+ "step": 260
264
+ },
265
+ {
266
+ "epoch": 8.852459016393443,
267
+ "grad_norm": 4.1779608726501465,
268
+ "learning_rate": 5.555555555555556e-06,
269
+ "loss": 0.2139,
270
+ "step": 270
271
  },
272
  {
273
+ "epoch": 8.98360655737705,
274
+ "eval_accuracy": 0.878345498783455,
275
+ "eval_loss": 0.2632748484611511,
276
+ "eval_runtime": 46.7653,
277
+ "eval_samples_per_second": 8.789,
278
+ "eval_steps_per_second": 0.15,
279
+ "step": 274
280
+ },
281
+ {
282
+ "epoch": 9.180327868852459,
283
+ "grad_norm": 2.4798853397369385,
284
+ "learning_rate": 3.7037037037037037e-06,
285
+ "loss": 0.2365,
286
+ "step": 280
287
+ },
288
+ {
289
+ "epoch": 9.508196721311476,
290
+ "grad_norm": 3.2969157695770264,
291
+ "learning_rate": 1.8518518518518519e-06,
292
+ "loss": 0.2242,
293
+ "step": 290
294
+ },
295
+ {
296
+ "epoch": 9.836065573770492,
297
+ "grad_norm": 2.2613918781280518,
298
  "learning_rate": 0.0,
299
+ "loss": 0.2087,
300
+ "step": 300
301
  },
302
  {
303
+ "epoch": 9.836065573770492,
304
+ "eval_accuracy": 0.8807785888077859,
305
+ "eval_loss": 0.2599342465400696,
306
+ "eval_runtime": 47.0372,
307
+ "eval_samples_per_second": 8.738,
308
+ "eval_steps_per_second": 0.149,
309
+ "step": 300
310
  },
311
  {
312
+ "epoch": 9.836065573770492,
313
+ "step": 300,
314
+ "total_flos": 1.9063087726729052e+18,
315
+ "train_loss": 0.3090126927693685,
316
+ "train_runtime": 8089.4797,
317
+ "train_samples_per_second": 9.637,
318
+ "train_steps_per_second": 0.037
319
  }
320
  ],
321
  "logging_steps": 10,
322
+ "max_steps": 300,
323
  "num_input_tokens_seen": 0,
324
  "num_train_epochs": 10,
325
  "save_steps": 500,
 
335
  "attributes": {}
336
  }
337
  },
338
+ "total_flos": 1.9063087726729052e+18,
339
  "train_batch_size": 64,
340
  "trial_name": null,
341
  "trial_params": null