Alawy21 commited on
Commit
9dc7208
·
verified ·
1 Parent(s): 0b55500

Training in progress, step 1000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd96db79c4222be60ac97ed313bf003c743cde5a7a5c4b544170bfe9a38918ab
3
  size 2026568872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43a9360070c3bc7665329ef6cb9c18a8e89dccf07afbcce13c03aaa1b9135b86
3
  size 2026568872
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:836e1bb9ed04bb2bb4d304d7b680e054ec780785a9b452982b60c396c2ef8eab
3
  size 4053320058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2279e40a4763df41f0c4bffc85edce65759b20ec629561bfa04dd1975a601194
3
  size 4053320058
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f37c40ce327861a7ca13b719d3aa37510a143368b6e74358bdb14becb3899e1e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b6047bd8cc6f4cdf7c46dea47edb8e542435510070c6cd1e0a7d9ccf5fd7da
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8469d184647c045d6d08f5cb32f7cd8ed66023f5033c0b4e25a0f6783d7d6ba9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:272640d7b61b2124df30234bf0964b9c0109c8e0621189f0064bc693bd54a0bb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 1.3774104683195592,
6
  "eval_steps": 363,
7
- "global_step": 500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -94,6 +94,84 @@
94
  "learning_rate": 0.00019964161211349086,
95
  "loss": 3.9705,
96
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
97
  }
98
  ],
99
  "logging_steps": 50,
@@ -113,7 +191,7 @@
113
  "attributes": {}
114
  }
115
  },
116
- "total_flos": 8795704000512000.0,
117
  "train_batch_size": 1,
118
  "trial_name": null,
119
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.7548209366391183,
6
  "eval_steps": 363,
7
+ "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
94
  "learning_rate": 0.00019964161211349086,
95
  "loss": 3.9705,
96
  "step": 500
97
+ },
98
+ {
99
+ "epoch": 1.5151515151515151,
100
+ "grad_norm": 8.101067543029785,
101
+ "learning_rate": 0.00019956463117038584,
102
+ "loss": 3.9586,
103
+ "step": 550
104
+ },
105
+ {
106
+ "epoch": 1.6528925619834711,
107
+ "grad_norm": 5.826588153839111,
108
+ "learning_rate": 0.00019948018457833974,
109
+ "loss": 3.9575,
110
+ "step": 600
111
+ },
112
+ {
113
+ "epoch": 1.790633608815427,
114
+ "grad_norm": 6.841236114501953,
115
+ "learning_rate": 0.00019938827866940642,
116
+ "loss": 3.9663,
117
+ "step": 650
118
+ },
119
+ {
120
+ "epoch": 1.9283746556473829,
121
+ "grad_norm": 7.117630481719971,
122
+ "learning_rate": 0.0001992889203349613,
123
+ "loss": 3.9456,
124
+ "step": 700
125
+ },
126
+ {
127
+ "epoch": 2.0,
128
+ "eval_loss": 4.049391269683838,
129
+ "eval_runtime": 70.275,
130
+ "eval_samples_per_second": 6.916,
131
+ "eval_steps_per_second": 6.916,
132
+ "step": 726
133
+ },
134
+ {
135
+ "epoch": 2.0661157024793386,
136
+ "grad_norm": 6.624067783355713,
137
+ "learning_rate": 0.00019918211702518463,
138
+ "loss": 3.789,
139
+ "step": 750
140
+ },
141
+ {
142
+ "epoch": 2.203856749311295,
143
+ "grad_norm": 5.628150463104248,
144
+ "learning_rate": 0.00019906787674850275,
145
+ "loss": 3.6285,
146
+ "step": 800
147
+ },
148
+ {
149
+ "epoch": 2.3415977961432506,
150
+ "grad_norm": 5.53456974029541,
151
+ "learning_rate": 0.00019894620807098763,
152
+ "loss": 3.6531,
153
+ "step": 850
154
+ },
155
+ {
156
+ "epoch": 2.479338842975207,
157
+ "grad_norm": 4.971586227416992,
158
+ "learning_rate": 0.00019881712011571466,
159
+ "loss": 3.6592,
160
+ "step": 900
161
+ },
162
+ {
163
+ "epoch": 2.6170798898071626,
164
+ "grad_norm": 5.7469801902771,
165
+ "learning_rate": 0.0001986806225620785,
166
+ "loss": 3.6672,
167
+ "step": 950
168
+ },
169
+ {
170
+ "epoch": 2.7548209366391183,
171
+ "grad_norm": 4.165513038635254,
172
+ "learning_rate": 0.00019853672564506725,
173
+ "loss": 3.6785,
174
+ "step": 1000
175
  }
176
  ],
177
  "logging_steps": 50,
 
191
  "attributes": {}
192
  }
193
  },
194
+ "total_flos": 1.7591408001024e+16,
195
  "train_batch_size": 1,
196
  "trial_name": null,
197
  "trial_params": null