Dohahemdann commited on
Commit
26e6fb1
·
verified ·
1 Parent(s): d61c5cb

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0120a8149bd4c9c80c434ee1d50d17d11b5ff6a9a540930d8477b8330665d7c6
3
  size 75579856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00fce450168446a0c820c73cad4d2fcd438098032a9cda56a421a3ae4ee33e86
3
  size 75579856
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cf508cce684f141d23779498e2107e05fe761d23d480c911d36daef0444c4a1
3
  size 39015574
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eac4ff7de5471e6e64cda3c96fc59903ed4d494cf08311d229e77212f103cdc5
3
  size 39015574
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:923d278f6956e6477248046106bcc976164d13b97aca384b8661e5bb580c9f50
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77653f901a737c32c712d4d98c59bda07889be56e9af169aa9525194aa1f203a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8ccbb7e9b2bee49a3a83a2cbe68b9ea55fce3c23d9b7073b07177a5b6c6d109
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b240dd9ddd423073bd07780b2cb840e4271f59c698cb81fac8efa021107c82d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,222 +2,82 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.020750475531730934,
6
  "eval_steps": 500,
7
- "global_step": 30,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.0006916825177243646,
14
- "grad_norm": 0.23008033633232117,
15
  "learning_rate": 0.0,
16
- "loss": 2.4935,
17
  "step": 1
18
  },
19
  {
20
  "epoch": 0.0013833650354487291,
21
- "grad_norm": 0.21738959848880768,
22
  "learning_rate": 1.3698630136986302e-06,
23
- "loss": 2.7335,
24
  "step": 2
25
  },
26
  {
27
  "epoch": 0.0020750475531730937,
28
- "grad_norm": 0.24479049444198608,
29
  "learning_rate": 2.7397260273972604e-06,
30
- "loss": 2.9014,
31
  "step": 3
32
  },
33
  {
34
  "epoch": 0.0027667300708974583,
35
- "grad_norm": 0.26578858494758606,
36
  "learning_rate": 4.10958904109589e-06,
37
- "loss": 2.8301,
38
  "step": 4
39
  },
40
  {
41
  "epoch": 0.0034584125886218224,
42
- "grad_norm": 0.21829573810100555,
43
  "learning_rate": 5.479452054794521e-06,
44
- "loss": 2.4889,
45
  "step": 5
46
  },
47
  {
48
  "epoch": 0.004150095106346187,
49
- "grad_norm": 0.24539536237716675,
50
  "learning_rate": 6.849315068493151e-06,
51
- "loss": 2.0441,
52
  "step": 6
53
  },
54
  {
55
  "epoch": 0.0048417776240705515,
56
- "grad_norm": 0.3143893778324127,
57
  "learning_rate": 8.21917808219178e-06,
58
- "loss": 2.0242,
59
  "step": 7
60
  },
61
  {
62
  "epoch": 0.0055334601417949165,
63
- "grad_norm": 0.26045775413513184,
64
  "learning_rate": 9.589041095890411e-06,
65
- "loss": 2.0649,
66
  "step": 8
67
  },
68
  {
69
  "epoch": 0.006225142659519281,
70
- "grad_norm": 0.1994040161371231,
71
  "learning_rate": 1.0958904109589042e-05,
72
- "loss": 2.6361,
73
  "step": 9
74
  },
75
  {
76
  "epoch": 0.006916825177243645,
77
- "grad_norm": 0.43011218309402466,
78
  "learning_rate": 1.2328767123287671e-05,
79
- "loss": 2.86,
80
  "step": 10
81
- },
82
- {
83
- "epoch": 0.00760850769496801,
84
- "grad_norm": 0.3276166021823883,
85
- "learning_rate": 1.3698630136986302e-05,
86
- "loss": 2.5563,
87
- "step": 11
88
- },
89
- {
90
- "epoch": 0.008300190212692375,
91
- "grad_norm": 0.6293095350265503,
92
- "learning_rate": 1.5068493150684931e-05,
93
- "loss": 1.6394,
94
- "step": 12
95
- },
96
- {
97
- "epoch": 0.008991872730416739,
98
- "grad_norm": 0.3222309648990631,
99
- "learning_rate": 1.643835616438356e-05,
100
- "loss": 1.8561,
101
- "step": 13
102
- },
103
- {
104
- "epoch": 0.009683555248141103,
105
- "grad_norm": 0.2540973126888275,
106
- "learning_rate": 1.780821917808219e-05,
107
- "loss": 3.0327,
108
- "step": 14
109
- },
110
- {
111
- "epoch": 0.010375237765865467,
112
- "grad_norm": 0.2383764237165451,
113
- "learning_rate": 1.9178082191780822e-05,
114
- "loss": 2.2992,
115
- "step": 15
116
- },
117
- {
118
- "epoch": 0.011066920283589833,
119
- "grad_norm": 0.35212621092796326,
120
- "learning_rate": 2.0547945205479453e-05,
121
- "loss": 2.8462,
122
- "step": 16
123
- },
124
- {
125
- "epoch": 0.011758602801314197,
126
- "grad_norm": 0.23731686174869537,
127
- "learning_rate": 2.1917808219178083e-05,
128
- "loss": 2.8213,
129
- "step": 17
130
- },
131
- {
132
- "epoch": 0.012450285319038561,
133
- "grad_norm": 0.3209351599216461,
134
- "learning_rate": 2.328767123287671e-05,
135
- "loss": 2.4589,
136
- "step": 18
137
- },
138
- {
139
- "epoch": 0.013141967836762926,
140
- "grad_norm": 0.22553832828998566,
141
- "learning_rate": 2.4657534246575342e-05,
142
- "loss": 2.358,
143
- "step": 19
144
- },
145
- {
146
- "epoch": 0.01383365035448729,
147
- "grad_norm": 0.29168522357940674,
148
- "learning_rate": 2.6027397260273973e-05,
149
- "loss": 2.5836,
150
- "step": 20
151
- },
152
- {
153
- "epoch": 0.014525332872211656,
154
- "grad_norm": 0.2680312991142273,
155
- "learning_rate": 2.7397260273972603e-05,
156
- "loss": 2.748,
157
- "step": 21
158
- },
159
- {
160
- "epoch": 0.01521701538993602,
161
- "grad_norm": 0.2818455398082733,
162
- "learning_rate": 2.8767123287671234e-05,
163
- "loss": 2.5701,
164
- "step": 22
165
- },
166
- {
167
- "epoch": 0.015908697907660384,
168
- "grad_norm": 0.2324114739894867,
169
- "learning_rate": 3.0136986301369862e-05,
170
- "loss": 1.8337,
171
- "step": 23
172
- },
173
- {
174
- "epoch": 0.01660038042538475,
175
- "grad_norm": 0.2770012617111206,
176
- "learning_rate": 3.1506849315068496e-05,
177
- "loss": 2.7497,
178
- "step": 24
179
- },
180
- {
181
- "epoch": 0.017292062943109112,
182
- "grad_norm": 0.27101394534111023,
183
- "learning_rate": 3.287671232876712e-05,
184
- "loss": 2.8125,
185
- "step": 25
186
- },
187
- {
188
- "epoch": 0.017983745460833478,
189
- "grad_norm": 0.27373552322387695,
190
- "learning_rate": 3.424657534246575e-05,
191
- "loss": 2.7676,
192
- "step": 26
193
- },
194
- {
195
- "epoch": 0.01867542797855784,
196
- "grad_norm": 0.2803725302219391,
197
- "learning_rate": 3.561643835616438e-05,
198
- "loss": 2.2638,
199
- "step": 27
200
- },
201
- {
202
- "epoch": 0.019367110496282206,
203
- "grad_norm": 0.25683945417404175,
204
- "learning_rate": 3.698630136986301e-05,
205
- "loss": 2.8515,
206
- "step": 28
207
- },
208
- {
209
- "epoch": 0.020058793014006572,
210
- "grad_norm": 0.32016420364379883,
211
- "learning_rate": 3.8356164383561644e-05,
212
- "loss": 2.8561,
213
- "step": 29
214
- },
215
- {
216
- "epoch": 0.020750475531730934,
217
- "grad_norm": 0.4947223663330078,
218
- "learning_rate": 3.9726027397260274e-05,
219
- "loss": 2.9034,
220
- "step": 30
221
  }
222
  ],
223
  "logging_steps": 1,
@@ -237,7 +97,7 @@
237
  "attributes": {}
238
  }
239
  },
240
- "total_flos": 1008346610073600.0,
241
  "train_batch_size": 2,
242
  "trial_name": null,
243
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.006916825177243645,
6
  "eval_steps": 500,
7
+ "global_step": 10,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.0006916825177243646,
14
+ "grad_norm": 0.2700531780719757,
15
  "learning_rate": 0.0,
16
+ "loss": 2.4618,
17
  "step": 1
18
  },
19
  {
20
  "epoch": 0.0013833650354487291,
21
+ "grad_norm": 0.26140472292900085,
22
  "learning_rate": 1.3698630136986302e-06,
23
+ "loss": 2.6975,
24
  "step": 2
25
  },
26
  {
27
  "epoch": 0.0020750475531730937,
28
+ "grad_norm": 0.29990413784980774,
29
  "learning_rate": 2.7397260273972604e-06,
30
+ "loss": 2.8554,
31
  "step": 3
32
  },
33
  {
34
  "epoch": 0.0027667300708974583,
35
+ "grad_norm": 0.31456881761550903,
36
  "learning_rate": 4.10958904109589e-06,
37
+ "loss": 2.7952,
38
  "step": 4
39
  },
40
  {
41
  "epoch": 0.0034584125886218224,
42
+ "grad_norm": 0.25708264112472534,
43
  "learning_rate": 5.479452054794521e-06,
44
+ "loss": 2.4612,
45
  "step": 5
46
  },
47
  {
48
  "epoch": 0.004150095106346187,
49
+ "grad_norm": 0.2855897843837738,
50
  "learning_rate": 6.849315068493151e-06,
51
+ "loss": 2.0206,
52
  "step": 6
53
  },
54
  {
55
  "epoch": 0.0048417776240705515,
56
+ "grad_norm": 0.35951822996139526,
57
  "learning_rate": 8.21917808219178e-06,
58
+ "loss": 1.9767,
59
  "step": 7
60
  },
61
  {
62
  "epoch": 0.0055334601417949165,
63
+ "grad_norm": 0.29928985238075256,
64
  "learning_rate": 9.589041095890411e-06,
65
+ "loss": 2.045,
66
  "step": 8
67
  },
68
  {
69
  "epoch": 0.006225142659519281,
70
+ "grad_norm": 0.2297569364309311,
71
  "learning_rate": 1.0958904109589042e-05,
72
+ "loss": 2.6072,
73
  "step": 9
74
  },
75
  {
76
  "epoch": 0.006916825177243645,
77
+ "grad_norm": 0.4355849623680115,
78
  "learning_rate": 1.2328767123287671e-05,
79
+ "loss": 2.7934,
80
  "step": 10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
81
  }
82
  ],
83
  "logging_steps": 1,
 
97
  "attributes": {}
98
  }
99
  },
100
+ "total_flos": 347472609361920.0,
101
  "train_batch_size": 2,
102
  "trial_name": null,
103
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:30bce305e780d7aa24e0d46ea3202c9b9e4aeefe52ad17f13b6285abef6f0170
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317972894c2f1ae307c137efb293243ed18c91dbff44283521f155311cd9e980
3
  size 5432