LushoAp commited on
Commit
3f55f6f
·
verified ·
1 Parent(s): bc39dc0

End of training

Browse files
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ base_model: MCG-NJU/videomae-base
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ model-index:
9
+ - name: videomae-base-finetuned-ucf101-subset
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # videomae-base-finetuned-ucf101-subset
17
+
18
+ This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.3131
21
+ - Accuracy: 0.9161
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 5e-05
41
+ - train_batch_size: 4
42
+ - eval_batch_size: 4
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_ratio: 0.1
47
+ - training_steps: 300
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | 1.6397 | 0.25 | 75 | 1.5641 | 0.4 |
54
+ | 0.8215 | 1.25 | 150 | 0.8054 | 0.6429 |
55
+ | 0.2651 | 2.25 | 225 | 0.4286 | 0.8429 |
56
+ | 0.202 | 3.25 | 300 | 0.3048 | 0.9 |
57
+
58
+
59
+ ### Framework versions
60
+
61
+ - Transformers 4.41.2
62
+ - Pytorch 2.3.1+cu121
63
+ - Datasets 2.20.0
64
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.25,
3
+ "eval_accuracy": 0.9161290322580645,
4
+ "eval_loss": 0.3130556344985962,
5
+ "eval_runtime": 44.5775,
6
+ "eval_samples_per_second": 3.477,
7
+ "eval_steps_per_second": 0.875
8
+ }
config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "MCG-NJU/videomae-base",
3
+ "architectures": [
4
+ "VideoMAEForVideoClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "decoder_hidden_size": 384,
8
+ "decoder_intermediate_size": 1536,
9
+ "decoder_num_attention_heads": 6,
10
+ "decoder_num_hidden_layers": 4,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.0,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "ApplyEyeMakeup",
16
+ "1": "ApplyLipstick",
17
+ "2": "Archery",
18
+ "3": "BabyCrawling",
19
+ "4": "BalanceBeam",
20
+ "5": "BandMarching",
21
+ "6": "BaseballPitch",
22
+ "7": "Basketball",
23
+ "8": "BasketballDunk",
24
+ "9": "BenchPress"
25
+ },
26
+ "image_size": 224,
27
+ "initializer_range": 0.02,
28
+ "intermediate_size": 3072,
29
+ "label2id": {
30
+ "ApplyEyeMakeup": 0,
31
+ "ApplyLipstick": 1,
32
+ "Archery": 2,
33
+ "BabyCrawling": 3,
34
+ "BalanceBeam": 4,
35
+ "BandMarching": 5,
36
+ "BaseballPitch": 6,
37
+ "Basketball": 7,
38
+ "BasketballDunk": 8,
39
+ "BenchPress": 9
40
+ },
41
+ "layer_norm_eps": 1e-12,
42
+ "model_type": "videomae",
43
+ "norm_pix_loss": true,
44
+ "num_attention_heads": 12,
45
+ "num_channels": 3,
46
+ "num_frames": 16,
47
+ "num_hidden_layers": 12,
48
+ "patch_size": 16,
49
+ "problem_type": "single_label_classification",
50
+ "qkv_bias": true,
51
+ "torch_dtype": "float32",
52
+ "transformers_version": "4.41.2",
53
+ "tubelet_size": 2,
54
+ "use_mean_pooling": false
55
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6e5f552194670320217649faf08e0059de358c41b5160a46b3c9ae254b76b4
3
+ size 344961984
preprocessor_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "videos",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_center_crop",
8
+ "crop_size",
9
+ "do_rescale",
10
+ "rescale_factor",
11
+ "do_normalize",
12
+ "image_mean",
13
+ "image_std",
14
+ "return_tensors",
15
+ "data_format",
16
+ "input_data_format"
17
+ ],
18
+ "crop_size": {
19
+ "height": 224,
20
+ "width": 224
21
+ },
22
+ "do_center_crop": true,
23
+ "do_normalize": true,
24
+ "do_rescale": true,
25
+ "do_resize": true,
26
+ "image_mean": [
27
+ 0.485,
28
+ 0.456,
29
+ 0.406
30
+ ],
31
+ "image_processor_type": "VideoMAEImageProcessor",
32
+ "image_std": [
33
+ 0.229,
34
+ 0.224,
35
+ 0.225
36
+ ],
37
+ "resample": 2,
38
+ "rescale_factor": 0.00392156862745098,
39
+ "size": {
40
+ "shortest_edge": 224
41
+ }
42
+ }
runs/Jun22_22-48-22_35076ae28850/events.out.tfevents.1719096505.35076ae28850.3689.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8e8f031b9953ddf144ed9e7e133b143ae168578359e6d8e6d6e3712dcb85612
3
+ size 16348
runs/Jun22_22-50-44_35076ae28850/events.out.tfevents.1719096665.35076ae28850.3689.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19f35a29adb96813698ef5020dc762c44d4c843bc84c0ae24ef030c74ef1559e
3
+ size 5507
runs/Jun22_22-51-19_35076ae28850/events.out.tfevents.1719096690.35076ae28850.3689.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55b514b99afc12c1cdfc806cb0dbfd3984496c3d5782e4539dbc481798126981
3
+ size 13429
runs/Jun22_22-51-19_35076ae28850/events.out.tfevents.1719100722.35076ae28850.3689.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06ddf6c60190867ab75076beef5c938d1909383400cdb72b9bf9f9fe2b82f456
3
+ size 734
test_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.25,
3
+ "eval_accuracy": 0.9161290322580645,
4
+ "eval_loss": 0.3130556344985962,
5
+ "eval_runtime": 44.5775,
6
+ "eval_samples_per_second": 3.477,
7
+ "eval_steps_per_second": 0.875
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,306 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9,
3
+ "best_model_checkpoint": "videomae-base-finetuned-ucf101-subset/checkpoint-300",
4
+ "epoch": 3.25,
5
+ "eval_steps": 500,
6
+ "global_step": 300,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.03333333333333333,
13
+ "grad_norm": 20.201610565185547,
14
+ "learning_rate": 1.6666666666666667e-05,
15
+ "loss": 2.444,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 0.06666666666666667,
20
+ "grad_norm": 12.654314994812012,
21
+ "learning_rate": 3.3333333333333335e-05,
22
+ "loss": 2.3019,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.1,
27
+ "grad_norm": 12.947044372558594,
28
+ "learning_rate": 5e-05,
29
+ "loss": 2.2352,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.13333333333333333,
34
+ "grad_norm": 11.306678771972656,
35
+ "learning_rate": 4.814814814814815e-05,
36
+ "loss": 2.1503,
37
+ "step": 40
38
+ },
39
+ {
40
+ "epoch": 0.16666666666666666,
41
+ "grad_norm": 11.73679256439209,
42
+ "learning_rate": 4.62962962962963e-05,
43
+ "loss": 2.0142,
44
+ "step": 50
45
+ },
46
+ {
47
+ "epoch": 0.2,
48
+ "grad_norm": 17.422954559326172,
49
+ "learning_rate": 4.4444444444444447e-05,
50
+ "loss": 1.9569,
51
+ "step": 60
52
+ },
53
+ {
54
+ "epoch": 0.23333333333333334,
55
+ "grad_norm": 12.51447868347168,
56
+ "learning_rate": 4.259259259259259e-05,
57
+ "loss": 1.6397,
58
+ "step": 70
59
+ },
60
+ {
61
+ "epoch": 0.25,
62
+ "eval_accuracy": 0.4,
63
+ "eval_loss": 1.5641058683395386,
64
+ "eval_runtime": 19.6018,
65
+ "eval_samples_per_second": 3.571,
66
+ "eval_steps_per_second": 0.918,
67
+ "step": 75
68
+ },
69
+ {
70
+ "epoch": 1.0166666666666666,
71
+ "grad_norm": 13.403763771057129,
72
+ "learning_rate": 4.074074074074074e-05,
73
+ "loss": 1.3891,
74
+ "step": 80
75
+ },
76
+ {
77
+ "epoch": 1.05,
78
+ "grad_norm": 11.36450481414795,
79
+ "learning_rate": 3.888888888888889e-05,
80
+ "loss": 0.9598,
81
+ "step": 90
82
+ },
83
+ {
84
+ "epoch": 1.0833333333333333,
85
+ "grad_norm": 14.846416473388672,
86
+ "learning_rate": 3.7037037037037037e-05,
87
+ "loss": 0.8928,
88
+ "step": 100
89
+ },
90
+ {
91
+ "epoch": 1.1166666666666667,
92
+ "grad_norm": 9.266874313354492,
93
+ "learning_rate": 3.518518518518519e-05,
94
+ "loss": 0.8165,
95
+ "step": 110
96
+ },
97
+ {
98
+ "epoch": 1.15,
99
+ "grad_norm": 21.095354080200195,
100
+ "learning_rate": 3.3333333333333335e-05,
101
+ "loss": 0.8177,
102
+ "step": 120
103
+ },
104
+ {
105
+ "epoch": 1.1833333333333333,
106
+ "grad_norm": 2.8057732582092285,
107
+ "learning_rate": 3.148148148148148e-05,
108
+ "loss": 0.6388,
109
+ "step": 130
110
+ },
111
+ {
112
+ "epoch": 1.2166666666666668,
113
+ "grad_norm": 7.6457672119140625,
114
+ "learning_rate": 2.962962962962963e-05,
115
+ "loss": 0.6413,
116
+ "step": 140
117
+ },
118
+ {
119
+ "epoch": 1.25,
120
+ "grad_norm": 15.42055892944336,
121
+ "learning_rate": 2.777777777777778e-05,
122
+ "loss": 0.8215,
123
+ "step": 150
124
+ },
125
+ {
126
+ "epoch": 1.25,
127
+ "eval_accuracy": 0.6428571428571429,
128
+ "eval_loss": 0.8053566813468933,
129
+ "eval_runtime": 20.071,
130
+ "eval_samples_per_second": 3.488,
131
+ "eval_steps_per_second": 0.897,
132
+ "step": 150
133
+ },
134
+ {
135
+ "epoch": 2.033333333333333,
136
+ "grad_norm": 30.937313079833984,
137
+ "learning_rate": 2.5925925925925925e-05,
138
+ "loss": 0.6518,
139
+ "step": 160
140
+ },
141
+ {
142
+ "epoch": 2.066666666666667,
143
+ "grad_norm": 3.2372474670410156,
144
+ "learning_rate": 2.4074074074074074e-05,
145
+ "loss": 0.5538,
146
+ "step": 170
147
+ },
148
+ {
149
+ "epoch": 2.1,
150
+ "grad_norm": 6.289671421051025,
151
+ "learning_rate": 2.2222222222222223e-05,
152
+ "loss": 0.3373,
153
+ "step": 180
154
+ },
155
+ {
156
+ "epoch": 2.1333333333333333,
157
+ "grad_norm": 28.10116195678711,
158
+ "learning_rate": 2.037037037037037e-05,
159
+ "loss": 0.4068,
160
+ "step": 190
161
+ },
162
+ {
163
+ "epoch": 2.1666666666666665,
164
+ "grad_norm": 2.922621726989746,
165
+ "learning_rate": 1.8518518518518518e-05,
166
+ "loss": 0.3649,
167
+ "step": 200
168
+ },
169
+ {
170
+ "epoch": 2.2,
171
+ "grad_norm": 1.8260467052459717,
172
+ "learning_rate": 1.6666666666666667e-05,
173
+ "loss": 0.4159,
174
+ "step": 210
175
+ },
176
+ {
177
+ "epoch": 2.2333333333333334,
178
+ "grad_norm": 7.952264308929443,
179
+ "learning_rate": 1.4814814814814815e-05,
180
+ "loss": 0.2651,
181
+ "step": 220
182
+ },
183
+ {
184
+ "epoch": 2.25,
185
+ "eval_accuracy": 0.8428571428571429,
186
+ "eval_loss": 0.4285981357097626,
187
+ "eval_runtime": 19.6403,
188
+ "eval_samples_per_second": 3.564,
189
+ "eval_steps_per_second": 0.916,
190
+ "step": 225
191
+ },
192
+ {
193
+ "epoch": 3.0166666666666666,
194
+ "grad_norm": 8.908254623413086,
195
+ "learning_rate": 1.2962962962962962e-05,
196
+ "loss": 0.3064,
197
+ "step": 230
198
+ },
199
+ {
200
+ "epoch": 3.05,
201
+ "grad_norm": 27.685514450073242,
202
+ "learning_rate": 1.1111111111111112e-05,
203
+ "loss": 0.2469,
204
+ "step": 240
205
+ },
206
+ {
207
+ "epoch": 3.0833333333333335,
208
+ "grad_norm": 6.450607776641846,
209
+ "learning_rate": 9.259259259259259e-06,
210
+ "loss": 0.31,
211
+ "step": 250
212
+ },
213
+ {
214
+ "epoch": 3.1166666666666667,
215
+ "grad_norm": 0.25113508105278015,
216
+ "learning_rate": 7.4074074074074075e-06,
217
+ "loss": 0.1282,
218
+ "step": 260
219
+ },
220
+ {
221
+ "epoch": 3.15,
222
+ "grad_norm": 4.858123779296875,
223
+ "learning_rate": 5.555555555555556e-06,
224
+ "loss": 0.1547,
225
+ "step": 270
226
+ },
227
+ {
228
+ "epoch": 3.183333333333333,
229
+ "grad_norm": 10.586645126342773,
230
+ "learning_rate": 3.7037037037037037e-06,
231
+ "loss": 0.1706,
232
+ "step": 280
233
+ },
234
+ {
235
+ "epoch": 3.216666666666667,
236
+ "grad_norm": 0.892095148563385,
237
+ "learning_rate": 1.8518518518518519e-06,
238
+ "loss": 0.2138,
239
+ "step": 290
240
+ },
241
+ {
242
+ "epoch": 3.25,
243
+ "grad_norm": 4.021522521972656,
244
+ "learning_rate": 0.0,
245
+ "loss": 0.202,
246
+ "step": 300
247
+ },
248
+ {
249
+ "epoch": 3.25,
250
+ "eval_accuracy": 0.9,
251
+ "eval_loss": 0.3048226535320282,
252
+ "eval_runtime": 19.4139,
253
+ "eval_samples_per_second": 3.606,
254
+ "eval_steps_per_second": 0.927,
255
+ "step": 300
256
+ },
257
+ {
258
+ "epoch": 3.25,
259
+ "step": 300,
260
+ "total_flos": 1.495384188125184e+18,
261
+ "train_loss": 0.8815905849138895,
262
+ "train_runtime": 827.7558,
263
+ "train_samples_per_second": 1.45,
264
+ "train_steps_per_second": 0.362
265
+ },
266
+ {
267
+ "epoch": 3.25,
268
+ "eval_accuracy": 0.9166666666666666,
269
+ "eval_loss": 0.31168052554130554,
270
+ "eval_runtime": 54.4422,
271
+ "eval_samples_per_second": 2.865,
272
+ "eval_steps_per_second": 0.716,
273
+ "step": 300
274
+ },
275
+ {
276
+ "epoch": 3.25,
277
+ "eval_accuracy": 0.9161290322580645,
278
+ "eval_loss": 0.3130556344985962,
279
+ "eval_runtime": 44.5775,
280
+ "eval_samples_per_second": 3.477,
281
+ "eval_steps_per_second": 0.875,
282
+ "step": 300
283
+ }
284
+ ],
285
+ "logging_steps": 10,
286
+ "max_steps": 300,
287
+ "num_input_tokens_seen": 0,
288
+ "num_train_epochs": 9223372036854775807,
289
+ "save_steps": 500,
290
+ "stateful_callbacks": {
291
+ "TrainerControl": {
292
+ "args": {
293
+ "should_epoch_stop": false,
294
+ "should_evaluate": false,
295
+ "should_log": false,
296
+ "should_save": true,
297
+ "should_training_stop": true
298
+ },
299
+ "attributes": {}
300
+ }
301
+ },
302
+ "total_flos": 1.495384188125184e+18,
303
+ "train_batch_size": 4,
304
+ "trial_name": null,
305
+ "trial_params": null
306
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a8a5b8456ba633c6bb1cab32266ed6c56d30e7211b126d1b0740b0bbff31fb4
3
+ size 5176