Shawon16 commited on
Commit
d95bd39
·
verified ·
1 Parent(s): 0604f38

End of training

Browse files
Files changed (4) hide show
  1. README.md +10 -8
  2. model.safetensors +1 -1
  3. trainer_state.json +88 -56
  4. training_args.bin +1 -1
README.md CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [MCG-NJU/videomae-base-finetuned-kinetics](https://huggingface.co/MCG-NJU/videomae-base-finetuned-kinetics) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.0029
22
  - Accuracy: 1.0
23
 
24
  ## Model description
@@ -54,13 +54,15 @@ The following hyperparameters were used during training:
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
56
  |:-------------:|:------:|:----:|:---------------:|:--------:|
57
- | 5.5194 | 0.02 | 7 | 1.2288 | 0.7143 |
58
- | 3.9958 | 1.0214 | 15 | 0.8649 | 1.0 |
59
- | 2.9836 | 2.02 | 22 | 0.5277 | 0.9286 |
60
- | 1.3684 | 3.0214 | 30 | 0.2611 | 1.0 |
61
- | 0.491 | 4.02 | 37 | 0.0589 | 1.0 |
62
- | 0.0773 | 5.0214 | 45 | 0.0096 | 1.0 |
63
- | 0.0159 | 6.02 | 52 | 0.0029 | 1.0 |
 
 
64
 
65
 
66
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [MCG-NJU/videomae-base-finetuned-kinetics](https://huggingface.co/MCG-NJU/videomae-base-finetuned-kinetics) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.0015
22
  - Accuracy: 1.0
23
 
24
  ## Model description
 
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
56
  |:-------------:|:------:|:----:|:---------------:|:--------:|
57
+ | 6.0723 | 0.02 | 7 | 1.3154 | 0.4286 |
58
+ | 4.5479 | 1.0214 | 15 | 1.0319 | 0.8571 |
59
+ | 3.6957 | 2.02 | 22 | 0.6454 | 0.8571 |
60
+ | 1.7212 | 3.0214 | 30 | 0.2796 | 1.0 |
61
+ | 0.6321 | 4.02 | 37 | 0.0587 | 1.0 |
62
+ | 0.0861 | 5.0214 | 45 | 0.0120 | 1.0 |
63
+ | 0.0166 | 6.02 | 52 | 0.0029 | 1.0 |
64
+ | 0.0047 | 7.0214 | 60 | 0.0018 | 1.0 |
65
+ | 0.0029 | 8.02 | 67 | 0.0015 | 1.0 |
66
 
67
 
68
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18755bed9fe08242090adee2cd3670f0c33c1c3714820d0ac8c3b7c39e1bdbdc
3
  size 344943488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:953acf335adb8764b7698ed56f60d19e918addc34c4386d841edeb0d032cfdbd
3
  size 344943488
trainer_state.json CHANGED
@@ -1,133 +1,165 @@
1
  {
2
  "best_metric": 1.0,
3
- "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/VideoMAE_4_CLASS_QUALITY_CHECK/checkpoint-15",
4
- "epoch": 6.02,
5
  "eval_steps": 500,
6
- "global_step": 52,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.02,
13
- "grad_norm": 18.13094711303711,
14
  "learning_rate": 1e-05,
15
- "loss": 5.5194,
16
  "step": 7
17
  },
18
  {
19
  "epoch": 0.02,
20
- "eval_accuracy": 0.7142857142857143,
21
- "eval_loss": 1.2288295030593872,
22
- "eval_runtime": 3.7225,
23
- "eval_samples_per_second": 3.761,
24
- "eval_steps_per_second": 1.88,
25
  "step": 7
26
  },
27
  {
28
  "epoch": 1.0214285714285714,
29
- "grad_norm": 24.789016723632812,
30
  "learning_rate": 2.1428571428571428e-05,
31
- "loss": 3.9958,
32
  "step": 15
33
  },
34
  {
35
  "epoch": 1.0214285714285714,
36
- "eval_accuracy": 1.0,
37
- "eval_loss": 0.8649204969406128,
38
- "eval_runtime": 2.6492,
39
- "eval_samples_per_second": 5.285,
40
- "eval_steps_per_second": 2.642,
41
  "step": 15
42
  },
43
  {
44
  "epoch": 2.02,
45
- "grad_norm": 27.01192283630371,
46
  "learning_rate": 3.142857142857143e-05,
47
- "loss": 2.9836,
48
  "step": 22
49
  },
50
  {
51
  "epoch": 2.02,
52
- "eval_accuracy": 0.9285714285714286,
53
- "eval_loss": 0.5277099609375,
54
- "eval_runtime": 3.371,
55
- "eval_samples_per_second": 4.153,
56
- "eval_steps_per_second": 2.077,
57
  "step": 22
58
  },
59
  {
60
  "epoch": 3.0214285714285714,
61
- "grad_norm": 19.976594924926758,
62
  "learning_rate": 4.1428571428571437e-05,
63
- "loss": 1.3684,
64
  "step": 30
65
  },
66
  {
67
  "epoch": 3.0214285714285714,
68
  "eval_accuracy": 1.0,
69
- "eval_loss": 0.2610691487789154,
70
- "eval_runtime": 2.7389,
71
- "eval_samples_per_second": 5.112,
72
- "eval_steps_per_second": 2.556,
73
  "step": 30
74
  },
75
  {
76
  "epoch": 4.02,
77
- "grad_norm": 3.972865104675293,
78
  "learning_rate": 4.9841269841269845e-05,
79
- "loss": 0.491,
80
  "step": 37
81
  },
82
  {
83
  "epoch": 4.02,
84
  "eval_accuracy": 1.0,
85
- "eval_loss": 0.0588662289083004,
86
- "eval_runtime": 2.4476,
87
- "eval_samples_per_second": 5.72,
88
- "eval_steps_per_second": 2.86,
89
  "step": 37
90
  },
91
  {
92
  "epoch": 5.021428571428571,
93
- "grad_norm": 0.7364888191223145,
94
  "learning_rate": 4.8571428571428576e-05,
95
- "loss": 0.0773,
96
  "step": 45
97
  },
98
  {
99
  "epoch": 5.021428571428571,
100
  "eval_accuracy": 1.0,
101
- "eval_loss": 0.00962775107473135,
102
- "eval_runtime": 3.1103,
103
- "eval_samples_per_second": 4.501,
104
- "eval_steps_per_second": 2.251,
105
  "step": 45
106
  },
107
  {
108
  "epoch": 6.02,
109
- "grad_norm": 0.1992603838443756,
110
  "learning_rate": 4.746031746031746e-05,
111
- "loss": 0.0159,
112
  "step": 52
113
  },
114
  {
115
  "epoch": 6.02,
116
  "eval_accuracy": 1.0,
117
- "eval_loss": 0.002915723016485572,
118
- "eval_runtime": 3.1585,
119
- "eval_samples_per_second": 4.432,
120
- "eval_steps_per_second": 2.216,
121
  "step": 52
122
  },
123
  {
124
- "epoch": 6.02,
125
- "step": 52,
126
- "total_flos": 5.2335646214455296e+17,
127
- "train_loss": 2.0500162014594445,
128
- "train_runtime": 126.4134,
129
- "train_samples_per_second": 22.15,
130
- "train_steps_per_second": 2.769
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  }
132
  ],
133
  "logging_steps": 500,
@@ -156,7 +188,7 @@
156
  "attributes": {}
157
  }
158
  },
159
- "total_flos": 5.2335646214455296e+17,
160
  "train_batch_size": 2,
161
  "trial_name": null,
162
  "trial_params": null
 
1
  {
2
  "best_metric": 1.0,
3
+ "best_model_checkpoint": "/media/cse/HDD/Shawon/shawon/10 fold timesformer/VideoMAE_4_CLASS_QUALITY_CHECK/checkpoint-30",
4
+ "epoch": 8.02,
5
  "eval_steps": 500,
6
+ "global_step": 67,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.02,
13
+ "grad_norm": 17.483346939086914,
14
  "learning_rate": 1e-05,
15
+ "loss": 6.0723,
16
  "step": 7
17
  },
18
  {
19
  "epoch": 0.02,
20
+ "eval_accuracy": 0.42857142857142855,
21
+ "eval_loss": 1.3153599500656128,
22
+ "eval_runtime": 2.0006,
23
+ "eval_samples_per_second": 6.998,
24
+ "eval_steps_per_second": 3.499,
25
  "step": 7
26
  },
27
  {
28
  "epoch": 1.0214285714285714,
29
+ "grad_norm": 20.505908966064453,
30
  "learning_rate": 2.1428571428571428e-05,
31
+ "loss": 4.5479,
32
  "step": 15
33
  },
34
  {
35
  "epoch": 1.0214285714285714,
36
+ "eval_accuracy": 0.8571428571428571,
37
+ "eval_loss": 1.0319474935531616,
38
+ "eval_runtime": 1.9993,
39
+ "eval_samples_per_second": 7.003,
40
+ "eval_steps_per_second": 3.501,
41
  "step": 15
42
  },
43
  {
44
  "epoch": 2.02,
45
+ "grad_norm": 27.64762306213379,
46
  "learning_rate": 3.142857142857143e-05,
47
+ "loss": 3.6957,
48
  "step": 22
49
  },
50
  {
51
  "epoch": 2.02,
52
+ "eval_accuracy": 0.8571428571428571,
53
+ "eval_loss": 0.6453508734703064,
54
+ "eval_runtime": 1.9837,
55
+ "eval_samples_per_second": 7.057,
56
+ "eval_steps_per_second": 3.529,
57
  "step": 22
58
  },
59
  {
60
  "epoch": 3.0214285714285714,
61
+ "grad_norm": 23.30771827697754,
62
  "learning_rate": 4.1428571428571437e-05,
63
+ "loss": 1.7212,
64
  "step": 30
65
  },
66
  {
67
  "epoch": 3.0214285714285714,
68
  "eval_accuracy": 1.0,
69
+ "eval_loss": 0.2796107828617096,
70
+ "eval_runtime": 1.9824,
71
+ "eval_samples_per_second": 7.062,
72
+ "eval_steps_per_second": 3.531,
73
  "step": 30
74
  },
75
  {
76
  "epoch": 4.02,
77
+ "grad_norm": 4.675124645233154,
78
  "learning_rate": 4.9841269841269845e-05,
79
+ "loss": 0.6321,
80
  "step": 37
81
  },
82
  {
83
  "epoch": 4.02,
84
  "eval_accuracy": 1.0,
85
+ "eval_loss": 0.05873652920126915,
86
+ "eval_runtime": 1.984,
87
+ "eval_samples_per_second": 7.056,
88
+ "eval_steps_per_second": 3.528,
89
  "step": 37
90
  },
91
  {
92
  "epoch": 5.021428571428571,
93
+ "grad_norm": 0.7413554787635803,
94
  "learning_rate": 4.8571428571428576e-05,
95
+ "loss": 0.0861,
96
  "step": 45
97
  },
98
  {
99
  "epoch": 5.021428571428571,
100
  "eval_accuracy": 1.0,
101
+ "eval_loss": 0.012020383961498737,
102
+ "eval_runtime": 1.9597,
103
+ "eval_samples_per_second": 7.144,
104
+ "eval_steps_per_second": 3.572,
105
  "step": 45
106
  },
107
  {
108
  "epoch": 6.02,
109
+ "grad_norm": 0.23269684612751007,
110
  "learning_rate": 4.746031746031746e-05,
111
+ "loss": 0.0166,
112
  "step": 52
113
  },
114
  {
115
  "epoch": 6.02,
116
  "eval_accuracy": 1.0,
117
+ "eval_loss": 0.002919810125604272,
118
+ "eval_runtime": 1.9438,
119
+ "eval_samples_per_second": 7.202,
120
+ "eval_steps_per_second": 3.601,
121
  "step": 52
122
  },
123
  {
124
+ "epoch": 7.021428571428571,
125
+ "grad_norm": 0.06729763746261597,
126
+ "learning_rate": 4.6190476190476194e-05,
127
+ "loss": 0.0047,
128
+ "step": 60
129
+ },
130
+ {
131
+ "epoch": 7.021428571428571,
132
+ "eval_accuracy": 1.0,
133
+ "eval_loss": 0.0018126964569091797,
134
+ "eval_runtime": 2.0381,
135
+ "eval_samples_per_second": 6.869,
136
+ "eval_steps_per_second": 3.435,
137
+ "step": 60
138
+ },
139
+ {
140
+ "epoch": 8.02,
141
+ "grad_norm": 0.040539324283599854,
142
+ "learning_rate": 4.507936507936508e-05,
143
+ "loss": 0.0029,
144
+ "step": 67
145
+ },
146
+ {
147
+ "epoch": 8.02,
148
+ "eval_accuracy": 1.0,
149
+ "eval_loss": 0.0014999593840911984,
150
+ "eval_runtime": 1.9648,
151
+ "eval_samples_per_second": 7.125,
152
+ "eval_steps_per_second": 3.563,
153
+ "step": 67
154
+ },
155
+ {
156
+ "epoch": 8.02,
157
+ "step": 67,
158
+ "total_flos": 6.728868799001395e+17,
159
+ "train_loss": 1.8480051827074877,
160
+ "train_runtime": 179.1547,
161
+ "train_samples_per_second": 15.629,
162
+ "train_steps_per_second": 1.954
163
  }
164
  ],
165
  "logging_steps": 500,
 
188
  "attributes": {}
189
  }
190
  },
191
+ "total_flos": 6.728868799001395e+17,
192
  "train_batch_size": 2,
193
  "trial_name": null,
194
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a013d9572411196384a99102848a4781f0c5748dcee7d41055d25cd9caa1f442
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09d22f455848886a768bd8daad7d2fa4c29ab567a5a5a53ad4ccc1835846b9f
3
  size 5304