dyllanesl commited on
Commit
13e4f09
·
verified ·
1 Parent(s): 29c5c15

Upload trainer_state.json

Browse files
Files changed (1) hide show
  1. trainer_state.json +192 -36
trainer_state.json CHANGED
@@ -1,82 +1,238 @@
1
  {
2
- "best_metric": null,
3
- "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
- "eval_steps": 500,
6
- "global_step": 99,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.30303030303030304,
13
- "grad_norm": 8.576181411743164,
14
- "learning_rate": 1.0000000000000002e-06,
15
- "loss": 0.2118,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.6060606060606061,
20
- "grad_norm": 0.2644074559211731,
21
- "learning_rate": 2.0000000000000003e-06,
22
- "loss": 0.0719,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.9090909090909091,
27
- "grad_norm": 0.2882717549800873,
28
- "learning_rate": 3e-06,
29
- "loss": 0.1005,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.2121212121212122,
34
- "grad_norm": 1.0808690786361694,
35
- "learning_rate": 4.000000000000001e-06,
36
- "loss": 0.0753,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 1.5151515151515151,
41
- "grad_norm": 0.14457939565181732,
42
- "learning_rate": 5e-06,
43
- "loss": 0.0294,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 1.8181818181818183,
48
- "grad_norm": 0.3328377306461334,
49
- "learning_rate": 6e-06,
50
- "loss": 0.0393,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 2.121212121212121,
55
- "grad_norm": 1.3251802921295166,
56
- "learning_rate": 7.000000000000001e-06,
57
- "loss": 0.0256,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 2.4242424242424243,
62
- "grad_norm": 0.18471165001392365,
63
- "learning_rate": 8.000000000000001e-06,
64
- "loss": 0.0071,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 2.7272727272727275,
69
- "grad_norm": 1.0178537368774414,
70
- "learning_rate": 9e-06,
71
- "loss": 0.0107,
72
  "step": 90
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73
  }
74
  ],
75
  "logging_steps": 10,
76
- "max_steps": 99,
77
  "num_input_tokens_seen": 0,
78
- "num_train_epochs": 3,
79
- "save_steps": 500,
80
  "stateful_callbacks": {
81
  "TrainerControl": {
82
  "args": {
@@ -89,7 +245,7 @@
89
  "attributes": {}
90
  }
91
  },
92
- "total_flos": 1.20913507233792e+17,
93
  "train_batch_size": 16,
94
  "trial_name": null,
95
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.0019312179647386074,
3
+ "best_model_checkpoint": "./vision_transformer_model_progress/checkpoint-200",
4
+ "epoch": 9.0,
5
+ "eval_steps": 100,
6
+ "global_step": 297,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.30303030303030304,
13
+ "grad_norm": 0.12155108898878098,
14
+ "learning_rate": 2.0000000000000003e-06,
15
+ "loss": 0.0158,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.6060606060606061,
20
+ "grad_norm": 1.0541080236434937,
21
+ "learning_rate": 4.000000000000001e-06,
22
+ "loss": 0.0479,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 0.9090909090909091,
27
+ "grad_norm": 0.03966180980205536,
28
+ "learning_rate": 6e-06,
29
+ "loss": 0.0047,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.2121212121212122,
34
+ "grad_norm": 0.06989104300737381,
35
+ "learning_rate": 8.000000000000001e-06,
36
+ "loss": 0.0045,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 1.5151515151515151,
41
+ "grad_norm": 0.0959181860089302,
42
+ "learning_rate": 1e-05,
43
+ "loss": 0.0027,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 1.8181818181818183,
48
+ "grad_norm": 0.05803044140338898,
49
+ "learning_rate": 1.2e-05,
50
+ "loss": 0.0281,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 2.121212121212121,
55
+ "grad_norm": 0.24542172253131866,
56
+ "learning_rate": 1.4000000000000001e-05,
57
+ "loss": 0.0049,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 2.4242424242424243,
62
+ "grad_norm": 0.0824781134724617,
63
+ "learning_rate": 1.6000000000000003e-05,
64
+ "loss": 0.0018,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 2.7272727272727275,
69
+ "grad_norm": 0.05576428771018982,
70
+ "learning_rate": 1.8e-05,
71
+ "loss": 0.0015,
72
  "step": 90
73
+ },
74
+ {
75
+ "epoch": 3.0303030303030303,
76
+ "grad_norm": 0.02234538644552231,
77
+ "learning_rate": 2e-05,
78
+ "loss": 0.0018,
79
+ "step": 100
80
+ },
81
+ {
82
+ "epoch": 3.0303030303030303,
83
+ "eval_loss": 0.0030471435748040676,
84
+ "eval_runtime": 12.9149,
85
+ "eval_samples_per_second": 10.066,
86
+ "eval_steps_per_second": 0.697,
87
+ "step": 100
88
+ },
89
+ {
90
+ "epoch": 3.3333333333333335,
91
+ "grad_norm": 0.014021256007254124,
92
+ "learning_rate": 2.2000000000000003e-05,
93
+ "loss": 0.001,
94
+ "step": 110
95
+ },
96
+ {
97
+ "epoch": 3.6363636363636362,
98
+ "grad_norm": 0.03260282799601555,
99
+ "learning_rate": 2.4e-05,
100
+ "loss": 0.0011,
101
+ "step": 120
102
+ },
103
+ {
104
+ "epoch": 3.9393939393939394,
105
+ "grad_norm": 0.020618194714188576,
106
+ "learning_rate": 2.6000000000000002e-05,
107
+ "loss": 0.0012,
108
+ "step": 130
109
+ },
110
+ {
111
+ "epoch": 4.242424242424242,
112
+ "grad_norm": 0.021722646430134773,
113
+ "learning_rate": 2.8000000000000003e-05,
114
+ "loss": 0.0009,
115
+ "step": 140
116
+ },
117
+ {
118
+ "epoch": 4.545454545454545,
119
+ "grad_norm": 0.015248863026499748,
120
+ "learning_rate": 3e-05,
121
+ "loss": 0.0007,
122
+ "step": 150
123
+ },
124
+ {
125
+ "epoch": 4.848484848484849,
126
+ "grad_norm": 0.00668317498639226,
127
+ "learning_rate": 3.2000000000000005e-05,
128
+ "loss": 0.0007,
129
+ "step": 160
130
+ },
131
+ {
132
+ "epoch": 5.151515151515151,
133
+ "grad_norm": 0.009767434559762478,
134
+ "learning_rate": 3.4000000000000007e-05,
135
+ "loss": 0.0006,
136
+ "step": 170
137
+ },
138
+ {
139
+ "epoch": 5.454545454545454,
140
+ "grad_norm": 0.0070835500955581665,
141
+ "learning_rate": 3.6e-05,
142
+ "loss": 0.0005,
143
+ "step": 180
144
+ },
145
+ {
146
+ "epoch": 5.757575757575758,
147
+ "grad_norm": 0.005514961667358875,
148
+ "learning_rate": 3.8e-05,
149
+ "loss": 0.0005,
150
+ "step": 190
151
+ },
152
+ {
153
+ "epoch": 6.0606060606060606,
154
+ "grad_norm": 0.007456360850483179,
155
+ "learning_rate": 4e-05,
156
+ "loss": 0.0005,
157
+ "step": 200
158
+ },
159
+ {
160
+ "epoch": 6.0606060606060606,
161
+ "eval_loss": 0.0019312179647386074,
162
+ "eval_runtime": 13.4333,
163
+ "eval_samples_per_second": 9.677,
164
+ "eval_steps_per_second": 0.67,
165
+ "step": 200
166
+ },
167
+ {
168
+ "epoch": 6.363636363636363,
169
+ "grad_norm": 0.0049119917675852776,
170
+ "learning_rate": 4.2e-05,
171
+ "loss": 0.0004,
172
+ "step": 210
173
+ },
174
+ {
175
+ "epoch": 6.666666666666667,
176
+ "grad_norm": 0.0046443757601082325,
177
+ "learning_rate": 4.4000000000000006e-05,
178
+ "loss": 0.0004,
179
+ "step": 220
180
+ },
181
+ {
182
+ "epoch": 6.96969696969697,
183
+ "grad_norm": 0.006061806809157133,
184
+ "learning_rate": 4.600000000000001e-05,
185
+ "loss": 0.0004,
186
+ "step": 230
187
+ },
188
+ {
189
+ "epoch": 7.2727272727272725,
190
+ "grad_norm": 0.004031210206449032,
191
+ "learning_rate": 4.8e-05,
192
+ "loss": 0.0004,
193
+ "step": 240
194
+ },
195
+ {
196
+ "epoch": 7.575757575757576,
197
+ "grad_norm": 0.004312009084969759,
198
+ "learning_rate": 5e-05,
199
+ "loss": 0.0003,
200
+ "step": 250
201
+ },
202
+ {
203
+ "epoch": 7.878787878787879,
204
+ "grad_norm": 0.004320390522480011,
205
+ "learning_rate": 5.2000000000000004e-05,
206
+ "loss": 0.0003,
207
+ "step": 260
208
+ },
209
+ {
210
+ "epoch": 8.181818181818182,
211
+ "grad_norm": 0.0036691236309707165,
212
+ "learning_rate": 5.4000000000000005e-05,
213
+ "loss": 0.0003,
214
+ "step": 270
215
+ },
216
+ {
217
+ "epoch": 8.484848484848484,
218
+ "grad_norm": 0.0040830825455486774,
219
+ "learning_rate": 5.6000000000000006e-05,
220
+ "loss": 0.0003,
221
+ "step": 280
222
+ },
223
+ {
224
+ "epoch": 8.787878787878787,
225
+ "grad_norm": 0.003878690768033266,
226
+ "learning_rate": 5.8e-05,
227
+ "loss": 0.0002,
228
+ "step": 290
229
  }
230
  ],
231
  "logging_steps": 10,
232
+ "max_steps": 297,
233
  "num_input_tokens_seen": 0,
234
+ "num_train_epochs": 9,
235
+ "save_steps": 100,
236
  "stateful_callbacks": {
237
  "TrainerControl": {
238
  "args": {
 
245
  "attributes": {}
246
  }
247
  },
248
+ "total_flos": 3.62740521701376e+17,
249
  "train_batch_size": 16,
250
  "trial_name": null,
251
  "trial_params": null