DylanJHJ commited on
Commit
6030f03
·
1 Parent(s): b79dbd2

update ablation study models

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +93 -0
  2. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/config.json +45 -0
  3. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/model.safetensors +3 -0
  4. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/optimizer.pt +3 -0
  5. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_0.pth +3 -0
  6. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_1.pth +3 -0
  7. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_2.pth +3 -0
  8. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_3.pth +3 -0
  9. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/scheduler.pt +3 -0
  10. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/trainer_state.json +799 -0
  11. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/training_args.bin +3 -0
  12. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/config.json +45 -0
  13. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/model.safetensors +3 -0
  14. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/optimizer.pt +3 -0
  15. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_0.pth +3 -0
  16. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_1.pth +3 -0
  17. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_2.pth +3 -0
  18. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_3.pth +3 -0
  19. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/scheduler.pt +3 -0
  20. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/trainer_state.json +1564 -0
  21. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/training_args.bin +3 -0
  22. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/config.json +45 -0
  23. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/model.safetensors +3 -0
  24. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/optimizer.pt +3 -0
  25. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_0.pth +3 -0
  26. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_1.pth +3 -0
  27. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_2.pth +3 -0
  28. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_3.pth +3 -0
  29. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/scheduler.pt +3 -0
  30. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/trainer_state.json +2329 -0
  31. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/training_args.bin +3 -0
  32. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/config.json +45 -0
  33. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/model.safetensors +3 -0
  34. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/optimizer.pt +3 -0
  35. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_0.pth +3 -0
  36. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_1.pth +3 -0
  37. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_2.pth +3 -0
  38. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_3.pth +3 -0
  39. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/scheduler.pt +3 -0
  40. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/trainer_state.json +3094 -0
  41. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/training_args.bin +3 -0
  42. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/config.json +45 -0
  43. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/model.safetensors +3 -0
  44. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/optimizer.pt +3 -0
  45. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_0.pth +3 -0
  46. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_1.pth +3 -0
  47. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_2.pth +3 -0
  48. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_3.pth +3 -0
  49. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/scheduler.pt +3 -0
  50. modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/trainer_state.json +3859 -0
.gitattributes CHANGED
@@ -186,3 +186,96 @@ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4.512/checkpoint-50
186
  .git/lfs/objects/c9/29/c9292f8084df08f44633c95111d71cf5055e02e10995567cc733f2bef07d3113 filter=lfs diff=lfs merge=lfs -text
187
  .git/lfs/objects/7e/97/7e973922297f0c493940a90076a687d8de67be2aa6db71915dd86d5c61411430 filter=lfs diff=lfs merge=lfs -text
188
  modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4.512/optimizer.pt filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
186
  .git/lfs/objects/c9/29/c9292f8084df08f44633c95111d71cf5055e02e10995567cc733f2bef07d3113 filter=lfs diff=lfs merge=lfs -text
187
  .git/lfs/objects/7e/97/7e973922297f0c493940a90076a687d8de67be2aa6db71915dd86d5c61411430 filter=lfs diff=lfs merge=lfs -text
188
  modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4.512/optimizer.pt filter=lfs diff=lfs merge=lfs -text
189
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
190
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
191
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
192
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
193
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
194
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
195
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
196
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
197
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
198
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
199
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
200
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
201
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
202
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
203
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
204
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
205
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
206
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
207
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
208
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
209
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
210
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
211
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
212
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
213
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
214
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
215
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
216
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
217
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
218
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
219
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
220
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
221
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
222
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
223
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
224
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
225
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
226
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-15000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
227
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-15000/model.safetensors filter=lfs diff=lfs merge=lfs -text
228
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-20000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
229
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-20000/model.safetensors filter=lfs diff=lfs merge=lfs -text
230
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-10000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
231
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-10000/model.safetensors filter=lfs diff=lfs merge=lfs -text
232
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
233
+ modernbert-crux-researchy-flatten.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
234
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
235
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
236
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
237
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
238
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
239
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
240
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
241
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
242
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
243
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
244
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
245
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
246
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
247
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
248
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
249
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
250
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
251
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
252
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
253
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
254
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
255
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
256
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
257
+ modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
258
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
259
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
260
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
261
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
262
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
263
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
264
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
265
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
266
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
267
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
268
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
269
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
270
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs merge=lfs -text
271
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
272
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
273
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/model.safetensors filter=lfs diff=lfs merge=lfs -text
274
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
275
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
276
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
277
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
278
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
279
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
280
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-5000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
281
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4/checkpoint-5000/model.safetensors filter=lfs diff=lfs merge=lfs -text
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd56a1262b3564d49c1dc685e5e53ebad9fb0773925dd5a93b507f36dc9f7c7
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6de8ed923ca2086d0104e84c05300e12e9d0899e81e5aecff0ec9bd73b35052c
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3044a568f2c48f55afab6e63cf0c046d3586784e77e39b1af825abecfb2cdba8
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b50176e6a2e04d27f26849df53fc28c50b0357c3f1bc26ca6b6bb8848991144
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbda571d0d08add9abed7e3b3b11bb38b8d472a92fd8409d00660f44b646db10
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5ce74a6352b774e1ca5b1ca2e8b3e684a2cc29987576c968c710783e2b33305
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:053b181ac2db5ceae48310494587c492212456802136bc3852c62ccc7b9ff38c
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,799 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.7127583749109052,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007127583749109052,
79
+ "grad_norm": 11.9375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.5314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014255167498218105,
86
+ "grad_norm": 11.125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.4333,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021382751247327157,
93
+ "grad_norm": 22.75,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.6002,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.02851033499643621,
100
+ "grad_norm": 13.3125,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.5692,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03563791874554526,
107
+ "grad_norm": 14.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.2221,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042765502494654314,
114
+ "grad_norm": 12.4375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.3981,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04989308624376337,
121
+ "grad_norm": 11.375,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.3844,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05702066999287242,
128
+ "grad_norm": 12.375,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.6756,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06414825374198146,
135
+ "grad_norm": 9.8125,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3874,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07127583749109052,
142
+ "grad_norm": 11.5,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.3705,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07840342124019957,
149
+ "grad_norm": 10.0,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3897,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08553100498930863,
156
+ "grad_norm": 11.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.4046,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09265858873841767,
163
+ "grad_norm": 8.25,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1084,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09978617248752673,
170
+ "grad_norm": 7.96875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.1872,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10691375623663578,
177
+ "grad_norm": 10.8125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.4332,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11404133998574484,
184
+ "grad_norm": 12.8125,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2686,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12116892373485388,
191
+ "grad_norm": 12.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.9751,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12829650748396293,
198
+ "grad_norm": 9.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1097,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.13542409123307197,
205
+ "grad_norm": 10.6875,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 1.0446,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14255167498218105,
212
+ "grad_norm": 8.625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.2052,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.1496792587312901,
219
+ "grad_norm": 9.5625,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 1.0325,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15680684248039914,
226
+ "grad_norm": 9.1875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9825,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16393442622950818,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0758,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17106200997861726,
240
+ "grad_norm": 7.9375,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.076,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1781895937277263,
247
+ "grad_norm": 8.25,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0498,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18531717747683535,
254
+ "grad_norm": 7.9375,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9616,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.1924447612259444,
261
+ "grad_norm": 7.71875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 1.0782,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19957234497505347,
268
+ "grad_norm": 8.6875,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.9648,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.2066999287241625,
275
+ "grad_norm": 9.375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9675,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21382751247327156,
282
+ "grad_norm": 7.53125,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.9063,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2209550962223806,
289
+ "grad_norm": 6.96875,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.8631,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.22808267997148968,
296
+ "grad_norm": 7.28125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.9451,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23521026372059872,
303
+ "grad_norm": 8.9375,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 1.0807,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24233784746970777,
310
+ "grad_norm": 7.09375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9637,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.2494654312188168,
317
+ "grad_norm": 6.25,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9216,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25659301496792586,
324
+ "grad_norm": 6.34375,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.9588,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2637205987170349,
331
+ "grad_norm": 6.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8696,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.27084818246614395,
338
+ "grad_norm": 6.0625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.951,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.27797576621525305,
345
+ "grad_norm": 6.6875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.9381,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.2851033499643621,
352
+ "grad_norm": 7.3125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.9314,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29223093371347114,
359
+ "grad_norm": 5.84375,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9212,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2993585174625802,
366
+ "grad_norm": 6.90625,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.8897,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.30648610121168923,
373
+ "grad_norm": 5.46875,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.9813,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3136136849607983,
380
+ "grad_norm": 4.46875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.9023,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3207412687099073,
387
+ "grad_norm": 5.8125,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.882,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.32786885245901637,
394
+ "grad_norm": 6.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9943,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33499643620812547,
401
+ "grad_norm": 12.0,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8606,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.3421240199572345,
408
+ "grad_norm": 6.125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.9137,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34925160370634356,
415
+ "grad_norm": 6.25,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8677,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3563791874554526,
422
+ "grad_norm": 5.46875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.9161,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.36350677120456165,
429
+ "grad_norm": 6.125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.9176,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.3706343549536707,
436
+ "grad_norm": 6.34375,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.871,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.37776193870277974,
443
+ "grad_norm": 6.28125,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.9097,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.3848895224518888,
450
+ "grad_norm": 5.1875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8721,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.3920171062009979,
457
+ "grad_norm": 7.15625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.9282,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39914468995010693,
464
+ "grad_norm": 5.375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8213,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.406272273699216,
471
+ "grad_norm": 7.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8527,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.413399857448325,
478
+ "grad_norm": 5.75,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.944,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42052744119743407,
485
+ "grad_norm": 6.5625,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.9269,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.4276550249465431,
492
+ "grad_norm": 5.53125,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.8261,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43478260869565216,
499
+ "grad_norm": 6.4375,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 1.0049,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.4419101924447612,
506
+ "grad_norm": 6.65625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.9438,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44903777619387025,
513
+ "grad_norm": 5.9375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.9527,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.45616535994297935,
520
+ "grad_norm": 6.375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.8917,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.4632929436920884,
527
+ "grad_norm": 4.625,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8554,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47042052744119744,
534
+ "grad_norm": 5.59375,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.978,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.4775481111903065,
541
+ "grad_norm": 5.0625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.774,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48467569493941554,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8388,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4918032786885246,
555
+ "grad_norm": 4.8125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.8888,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4989308624376336,
562
+ "grad_norm": 6.03125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.7606,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5060584461867427,
569
+ "grad_norm": 5.1875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.9804,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5131860299358517,
576
+ "grad_norm": 6.5625,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.8832,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5203136136849608,
583
+ "grad_norm": 5.0,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.8385,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5274411974340698,
590
+ "grad_norm": 6.78125,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.9106,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5345687811831789,
597
+ "grad_norm": 6.3125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.807,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5416963649322879,
604
+ "grad_norm": 4.0,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7932,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5488239486813971,
611
+ "grad_norm": 5.53125,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.7533,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5559515324305061,
618
+ "grad_norm": 5.125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.8676,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5630791161796151,
625
+ "grad_norm": 5.0625,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8783,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5702066999287242,
632
+ "grad_norm": 5.5625,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.9352,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5773342836778332,
639
+ "grad_norm": 4.8125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8814,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5844618674269423,
646
+ "grad_norm": 5.96875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.8252,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5915894511760513,
653
+ "grad_norm": 4.96875,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8695,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5987170349251604,
660
+ "grad_norm": 5.21875,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8821,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6058446186742694,
667
+ "grad_norm": 6.375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.8968,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6129722024233785,
674
+ "grad_norm": 4.59375,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.9109,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6200997861724875,
681
+ "grad_norm": 6.625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.8701,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6272273699215966,
688
+ "grad_norm": 5.71875,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.8649,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6343549536707056,
695
+ "grad_norm": 5.25,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8578,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6414825374198146,
702
+ "grad_norm": 4.9375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.9067,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6486101211689237,
709
+ "grad_norm": 6.8125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7862,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6557377049180327,
716
+ "grad_norm": 5.875,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.8275,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6628652886671419,
723
+ "grad_norm": 5.6875,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.9155,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6699928724162509,
730
+ "grad_norm": 5.6875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.8526,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.67712045616536,
737
+ "grad_norm": 4.9375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8123,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.684248039914469,
744
+ "grad_norm": 5.4375,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7986,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6913756236635781,
751
+ "grad_norm": 5.6875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.9203,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6985032074126871,
758
+ "grad_norm": 5.78125,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7805,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7056307911617962,
765
+ "grad_norm": 5.4375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8683,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7127583749109052,
772
+ "grad_norm": 5.46875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8515,
775
+ "step": 1000
776
+ }
777
+ ],
778
+ "logging_steps": 10,
779
+ "max_steps": 5000,
780
+ "num_input_tokens_seen": 0,
781
+ "num_train_epochs": 4,
782
+ "save_steps": 1000,
783
+ "stateful_callbacks": {
784
+ "TrainerControl": {
785
+ "args": {
786
+ "should_epoch_stop": false,
787
+ "should_evaluate": false,
788
+ "should_log": false,
789
+ "should_save": true,
790
+ "should_training_stop": false
791
+ },
792
+ "attributes": {}
793
+ }
794
+ },
795
+ "total_flos": 0.0,
796
+ "train_batch_size": 16,
797
+ "trial_name": null,
798
+ "trial_params": null
799
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1afec67e7d18e7478623434fcf8c1bcc7719c01c4f7874089687d7138953afff
3
+ size 6097
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24ce79534477c2159ef6b96ac6d5331aa78395ddcb84d9ad63259b832e3f3b71
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83f07c64dc00bceb661a5db33fbba0361af657ed1cd530bc5850eb23fed589d7
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18bb6b6e7da6250d22c25b95c64287be4aea598d9d97ef67b0fd69b9be869ed7
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cff16289615ba210b601be2162794d1949cb1d62132099a4cf62330c43649a5
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6caa3980c911491b74ec2e96e48a78cd7c3365dd32116c61642a4b6839a2da1
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e93d0d6464a92d227e9235c134b2224cf28ea45ba375cf82e6949c632e2b0d5
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff65ca8b6e225d1d82cc63887ee464e8c896624115a4dc693eae1fff20815699
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/trainer_state.json ADDED
@@ -0,0 +1,1564 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.4255167498218104,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 2000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007127583749109052,
79
+ "grad_norm": 11.9375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.5314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014255167498218105,
86
+ "grad_norm": 11.125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.4333,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021382751247327157,
93
+ "grad_norm": 22.75,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.6002,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.02851033499643621,
100
+ "grad_norm": 13.3125,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.5692,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03563791874554526,
107
+ "grad_norm": 14.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.2221,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042765502494654314,
114
+ "grad_norm": 12.4375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.3981,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04989308624376337,
121
+ "grad_norm": 11.375,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.3844,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05702066999287242,
128
+ "grad_norm": 12.375,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.6756,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06414825374198146,
135
+ "grad_norm": 9.8125,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3874,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07127583749109052,
142
+ "grad_norm": 11.5,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.3705,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07840342124019957,
149
+ "grad_norm": 10.0,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3897,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08553100498930863,
156
+ "grad_norm": 11.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.4046,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09265858873841767,
163
+ "grad_norm": 8.25,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1084,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09978617248752673,
170
+ "grad_norm": 7.96875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.1872,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10691375623663578,
177
+ "grad_norm": 10.8125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.4332,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11404133998574484,
184
+ "grad_norm": 12.8125,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2686,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12116892373485388,
191
+ "grad_norm": 12.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.9751,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12829650748396293,
198
+ "grad_norm": 9.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1097,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.13542409123307197,
205
+ "grad_norm": 10.6875,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 1.0446,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14255167498218105,
212
+ "grad_norm": 8.625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.2052,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.1496792587312901,
219
+ "grad_norm": 9.5625,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 1.0325,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15680684248039914,
226
+ "grad_norm": 9.1875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9825,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16393442622950818,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0758,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17106200997861726,
240
+ "grad_norm": 7.9375,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.076,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1781895937277263,
247
+ "grad_norm": 8.25,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0498,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18531717747683535,
254
+ "grad_norm": 7.9375,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9616,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.1924447612259444,
261
+ "grad_norm": 7.71875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 1.0782,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19957234497505347,
268
+ "grad_norm": 8.6875,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.9648,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.2066999287241625,
275
+ "grad_norm": 9.375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9675,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21382751247327156,
282
+ "grad_norm": 7.53125,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.9063,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2209550962223806,
289
+ "grad_norm": 6.96875,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.8631,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.22808267997148968,
296
+ "grad_norm": 7.28125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.9451,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23521026372059872,
303
+ "grad_norm": 8.9375,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 1.0807,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24233784746970777,
310
+ "grad_norm": 7.09375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9637,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.2494654312188168,
317
+ "grad_norm": 6.25,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9216,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25659301496792586,
324
+ "grad_norm": 6.34375,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.9588,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2637205987170349,
331
+ "grad_norm": 6.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8696,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.27084818246614395,
338
+ "grad_norm": 6.0625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.951,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.27797576621525305,
345
+ "grad_norm": 6.6875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.9381,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.2851033499643621,
352
+ "grad_norm": 7.3125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.9314,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29223093371347114,
359
+ "grad_norm": 5.84375,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9212,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2993585174625802,
366
+ "grad_norm": 6.90625,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.8897,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.30648610121168923,
373
+ "grad_norm": 5.46875,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.9813,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3136136849607983,
380
+ "grad_norm": 4.46875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.9023,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3207412687099073,
387
+ "grad_norm": 5.8125,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.882,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.32786885245901637,
394
+ "grad_norm": 6.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9943,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33499643620812547,
401
+ "grad_norm": 12.0,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8606,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.3421240199572345,
408
+ "grad_norm": 6.125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.9137,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34925160370634356,
415
+ "grad_norm": 6.25,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8677,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3563791874554526,
422
+ "grad_norm": 5.46875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.9161,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.36350677120456165,
429
+ "grad_norm": 6.125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.9176,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.3706343549536707,
436
+ "grad_norm": 6.34375,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.871,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.37776193870277974,
443
+ "grad_norm": 6.28125,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.9097,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.3848895224518888,
450
+ "grad_norm": 5.1875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8721,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.3920171062009979,
457
+ "grad_norm": 7.15625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.9282,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39914468995010693,
464
+ "grad_norm": 5.375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8213,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.406272273699216,
471
+ "grad_norm": 7.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8527,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.413399857448325,
478
+ "grad_norm": 5.75,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.944,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42052744119743407,
485
+ "grad_norm": 6.5625,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.9269,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.4276550249465431,
492
+ "grad_norm": 5.53125,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.8261,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43478260869565216,
499
+ "grad_norm": 6.4375,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 1.0049,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.4419101924447612,
506
+ "grad_norm": 6.65625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.9438,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44903777619387025,
513
+ "grad_norm": 5.9375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.9527,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.45616535994297935,
520
+ "grad_norm": 6.375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.8917,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.4632929436920884,
527
+ "grad_norm": 4.625,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8554,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47042052744119744,
534
+ "grad_norm": 5.59375,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.978,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.4775481111903065,
541
+ "grad_norm": 5.0625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.774,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48467569493941554,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8388,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4918032786885246,
555
+ "grad_norm": 4.8125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.8888,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4989308624376336,
562
+ "grad_norm": 6.03125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.7606,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5060584461867427,
569
+ "grad_norm": 5.1875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.9804,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5131860299358517,
576
+ "grad_norm": 6.5625,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.8832,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5203136136849608,
583
+ "grad_norm": 5.0,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.8385,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5274411974340698,
590
+ "grad_norm": 6.78125,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.9106,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5345687811831789,
597
+ "grad_norm": 6.3125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.807,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5416963649322879,
604
+ "grad_norm": 4.0,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7932,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5488239486813971,
611
+ "grad_norm": 5.53125,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.7533,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5559515324305061,
618
+ "grad_norm": 5.125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.8676,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5630791161796151,
625
+ "grad_norm": 5.0625,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8783,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5702066999287242,
632
+ "grad_norm": 5.5625,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.9352,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5773342836778332,
639
+ "grad_norm": 4.8125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8814,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5844618674269423,
646
+ "grad_norm": 5.96875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.8252,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5915894511760513,
653
+ "grad_norm": 4.96875,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8695,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5987170349251604,
660
+ "grad_norm": 5.21875,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8821,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6058446186742694,
667
+ "grad_norm": 6.375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.8968,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6129722024233785,
674
+ "grad_norm": 4.59375,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.9109,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6200997861724875,
681
+ "grad_norm": 6.625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.8701,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6272273699215966,
688
+ "grad_norm": 5.71875,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.8649,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6343549536707056,
695
+ "grad_norm": 5.25,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8578,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6414825374198146,
702
+ "grad_norm": 4.9375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.9067,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6486101211689237,
709
+ "grad_norm": 6.8125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7862,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6557377049180327,
716
+ "grad_norm": 5.875,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.8275,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6628652886671419,
723
+ "grad_norm": 5.6875,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.9155,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6699928724162509,
730
+ "grad_norm": 5.6875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.8526,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.67712045616536,
737
+ "grad_norm": 4.9375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8123,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.684248039914469,
744
+ "grad_norm": 5.4375,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7986,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6913756236635781,
751
+ "grad_norm": 5.6875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.9203,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6985032074126871,
758
+ "grad_norm": 5.78125,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7805,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7056307911617962,
765
+ "grad_norm": 5.4375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8683,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7127583749109052,
772
+ "grad_norm": 5.46875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8515,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7127583749109052,
779
+ "eval/nano_beir.arguana": 0.49904762134680597,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7127583749109052,
784
+ "eval/nano_beir.climate_fever": 0.2710368630075979,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7127583749109052,
789
+ "eval/nano_beir.dbpedia_entity": 0.32627448935866016,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7127583749109052,
794
+ "eval/nano_beir.fever": 0.6251056136949217,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7127583749109052,
799
+ "eval/nano_beir.fiqa": 0.4264627976119544,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7127583749109052,
804
+ "eval/nano_beir.hotpotqa": 0.67890984670666,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7127583749109052,
809
+ "eval/nano_beir.nfcorpus": 0.33575253271342886,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7127583749109052,
814
+ "eval/nano_beir.nq": 0.41529049449791683,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7127583749109052,
819
+ "eval/nano_beir.quora": 0.8400060722791932,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7127583749109052,
824
+ "eval/nano_beir.scidocs": 0.25236583848912914,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7127583749109052,
829
+ "eval/nano_beir.scifact": 0.7248783076751154,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7127583749109052,
834
+ "eval/nano_beir.webis_touche2020": 0.29804501841705877,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7127583749109052,
839
+ "eval/avg": 0.4744312913165369,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7198859586600143,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8998,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7270135424091233,
851
+ "grad_norm": 5.34375,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8763,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7341411261582323,
858
+ "grad_norm": 4.875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8851,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7412687099073414,
865
+ "grad_norm": 4.40625,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.8247,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7483962936564504,
872
+ "grad_norm": 6.5625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.742,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7555238774055595,
879
+ "grad_norm": 4.6875,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8707,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7626514611546685,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7698,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7697790449037776,
893
+ "grad_norm": 6.1875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.9318,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7769066286528866,
900
+ "grad_norm": 4.40625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.752,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7840342124019958,
907
+ "grad_norm": 6.0,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.9631,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7911617961511048,
914
+ "grad_norm": 4.71875,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.8623,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7982893799002139,
921
+ "grad_norm": 5.84375,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8677,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8054169636493229,
928
+ "grad_norm": 4.03125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.7795,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.812544547398432,
935
+ "grad_norm": 4.125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.8011,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.819672131147541,
942
+ "grad_norm": 4.875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7984,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.82679971489665,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.8049,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8339272986457591,
956
+ "grad_norm": 5.15625,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7839,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8410548823948681,
963
+ "grad_norm": 4.78125,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7809,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8481824661439772,
970
+ "grad_norm": 4.71875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.7688,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8553100498930862,
977
+ "grad_norm": 7.75,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.8638,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8624376336421953,
984
+ "grad_norm": 5.9375,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8409,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8695652173913043,
991
+ "grad_norm": 5.875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.916,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8766928011404134,
998
+ "grad_norm": 5.40625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.9114,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.8838203848895224,
1005
+ "grad_norm": 4.5,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.8744,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.8909479686386315,
1012
+ "grad_norm": 4.9375,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.8471,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8980755523877405,
1019
+ "grad_norm": 5.21875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8014,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9052031361368497,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.8587,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.9123307198859587,
1033
+ "grad_norm": 5.9375,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8003,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.9194583036350678,
1040
+ "grad_norm": 6.09375,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7404,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.9265858873841768,
1047
+ "grad_norm": 5.5,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.722,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9337134711332858,
1054
+ "grad_norm": 5.21875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7005,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9408410548823949,
1061
+ "grad_norm": 6.78125,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.7987,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.9479686386315039,
1068
+ "grad_norm": 5.78125,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7872,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.955096222380613,
1075
+ "grad_norm": 4.875,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.8632,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.962223806129722,
1082
+ "grad_norm": 6.96875,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.8165,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9693513898788311,
1089
+ "grad_norm": 5.0625,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7802,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9764789736279401,
1096
+ "grad_norm": 5.34375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.8668,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9836065573770492,
1103
+ "grad_norm": 5.84375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.7785,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9907341411261582,
1110
+ "grad_norm": 6.25,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8952,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9978617248752673,
1117
+ "grad_norm": 6.34375,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.826,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.0049893086243764,
1124
+ "grad_norm": 6.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.9168,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0121168923734853,
1131
+ "grad_norm": 5.625,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8506,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0192444761225945,
1138
+ "grad_norm": 4.875,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.9257,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0263720598717034,
1145
+ "grad_norm": 5.96875,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.8302,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0334996436208126,
1152
+ "grad_norm": 5.375,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.7898,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0406272273699215,
1159
+ "grad_norm": 5.75,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.7682,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0477548111190307,
1166
+ "grad_norm": 4.9375,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.8378,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0548823948681396,
1173
+ "grad_norm": 4.15625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.7113,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0620099786172488,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.7187,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0691375623663577,
1187
+ "grad_norm": 5.21875,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.7843,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0762651461154669,
1194
+ "grad_norm": 5.28125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.7253,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0833927298645758,
1201
+ "grad_norm": 4.78125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.8077,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.090520313613685,
1208
+ "grad_norm": 5.8125,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.8465,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0976478973627941,
1215
+ "grad_norm": 4.875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.7467,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.104775481111903,
1222
+ "grad_norm": 5.0,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.7919,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1119030648610122,
1229
+ "grad_norm": 5.78125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7399,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1190306486101211,
1236
+ "grad_norm": 5.28125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.8305,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1261582323592303,
1243
+ "grad_norm": 5.09375,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.8986,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1332858161083392,
1250
+ "grad_norm": 5.65625,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8041,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1404133998574484,
1257
+ "grad_norm": 5.5625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8434,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1475409836065573,
1264
+ "grad_norm": 5.9375,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7681,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1546685673556665,
1271
+ "grad_norm": 4.71875,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.6835,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1617961511047754,
1278
+ "grad_norm": 4.8125,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.8917,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1689237348538846,
1285
+ "grad_norm": 5.25,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8296,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1760513186029935,
1292
+ "grad_norm": 4.90625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8147,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1831789023521027,
1299
+ "grad_norm": 6.90625,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.7656,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1903064861012118,
1306
+ "grad_norm": 5.28125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8931,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1974340698503207,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.7844,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2045616535994297,
1320
+ "grad_norm": 5.53125,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7334,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2116892373485388,
1327
+ "grad_norm": 5.28125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7509,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.218816821097648,
1334
+ "grad_norm": 5.5,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8287,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.225944404846757,
1341
+ "grad_norm": 5.4375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.8383,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.233071988595866,
1348
+ "grad_norm": 6.6875,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7443,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.240199572344975,
1355
+ "grad_norm": 6.0625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7868,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2473271560940842,
1362
+ "grad_norm": 5.09375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.7662,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2544547398431931,
1369
+ "grad_norm": 4.90625,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7558,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2615823235923023,
1376
+ "grad_norm": 5.34375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6894,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2687099073414112,
1383
+ "grad_norm": 5.15625,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7782,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2758374910905204,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.756,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2829650748396293,
1397
+ "grad_norm": 4.125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.763,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.2900926585887385,
1404
+ "grad_norm": 10.5,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7255,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.2972202423378474,
1411
+ "grad_norm": 4.65625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7615,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3043478260869565,
1418
+ "grad_norm": 5.15625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8092,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3114754098360657,
1425
+ "grad_norm": 4.71875,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7503,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3186029935851746,
1432
+ "grad_norm": 4.5,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.8121,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.3257305773342836,
1439
+ "grad_norm": 4.78125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.8773,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.3328581610833927,
1446
+ "grad_norm": 5.6875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.8202,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.3399857448325019,
1453
+ "grad_norm": 6.15625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.8359,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3471133285816108,
1460
+ "grad_norm": 5.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.9009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.3542409123307197,
1467
+ "grad_norm": 4.96875,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.7515,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.361368496079829,
1474
+ "grad_norm": 4.96875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.8321,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.368496079828938,
1481
+ "grad_norm": 6.875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.8211,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.375623663578047,
1488
+ "grad_norm": 3.921875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7751,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.3827512473271562,
1495
+ "grad_norm": 5.53125,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8174,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.389878831076265,
1502
+ "grad_norm": 5.28125,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7058,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3970064148253742,
1509
+ "grad_norm": 6.34375,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7263,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.4041339985744832,
1516
+ "grad_norm": 5.5625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.7503,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.4112615823235923,
1523
+ "grad_norm": 5.34375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.87,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4183891660727013,
1530
+ "grad_norm": 4.84375,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7866,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4255167498218104,
1537
+ "grad_norm": 5.0625,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.787,
1540
+ "step": 2000
1541
+ }
1542
+ ],
1543
+ "logging_steps": 10,
1544
+ "max_steps": 5000,
1545
+ "num_input_tokens_seen": 0,
1546
+ "num_train_epochs": 4,
1547
+ "save_steps": 1000,
1548
+ "stateful_callbacks": {
1549
+ "TrainerControl": {
1550
+ "args": {
1551
+ "should_epoch_stop": false,
1552
+ "should_evaluate": false,
1553
+ "should_log": false,
1554
+ "should_save": true,
1555
+ "should_training_stop": false
1556
+ },
1557
+ "attributes": {}
1558
+ }
1559
+ },
1560
+ "total_flos": 0.0,
1561
+ "train_batch_size": 16,
1562
+ "trial_name": null,
1563
+ "trial_params": null
1564
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-2000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1afec67e7d18e7478623434fcf8c1bcc7719c01c4f7874089687d7138953afff
3
+ size 6097
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0904555a863ff6d0ac55c9e1d292dce677e6a2180be364f66d90ff64f37d5dec
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fed561d644f364b4cc3ae8de2596c1b176c1c8a72072585e4390c0165e177921
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0e770c61e3acef7bd144efee65a450ed487b244ed293a8dd801394ffcae775
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21354c9980d6ba769d430eb9962ffba457eb04a66cb90e33b0934d3157fae7cf
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d574a8567b1636a399633d2b20273fdb3be4b888e37f0d331cd695b973f10463
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cca2b80df90b78b3ba9d9d2951c5b7aa80fc0a44b4512b495de6de40396fcb6e
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a8f419d963b829163d14296852a07ad341e5e63e49d44de7c0a10c9369d217
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/trainer_state.json ADDED
@@ -0,0 +1,2329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.1382751247327154,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 3000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007127583749109052,
79
+ "grad_norm": 11.9375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.5314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014255167498218105,
86
+ "grad_norm": 11.125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.4333,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021382751247327157,
93
+ "grad_norm": 22.75,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.6002,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.02851033499643621,
100
+ "grad_norm": 13.3125,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.5692,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03563791874554526,
107
+ "grad_norm": 14.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.2221,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042765502494654314,
114
+ "grad_norm": 12.4375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.3981,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04989308624376337,
121
+ "grad_norm": 11.375,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.3844,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05702066999287242,
128
+ "grad_norm": 12.375,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.6756,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06414825374198146,
135
+ "grad_norm": 9.8125,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3874,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07127583749109052,
142
+ "grad_norm": 11.5,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.3705,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07840342124019957,
149
+ "grad_norm": 10.0,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3897,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08553100498930863,
156
+ "grad_norm": 11.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.4046,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09265858873841767,
163
+ "grad_norm": 8.25,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1084,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09978617248752673,
170
+ "grad_norm": 7.96875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.1872,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10691375623663578,
177
+ "grad_norm": 10.8125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.4332,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11404133998574484,
184
+ "grad_norm": 12.8125,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2686,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12116892373485388,
191
+ "grad_norm": 12.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.9751,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12829650748396293,
198
+ "grad_norm": 9.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1097,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.13542409123307197,
205
+ "grad_norm": 10.6875,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 1.0446,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14255167498218105,
212
+ "grad_norm": 8.625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.2052,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.1496792587312901,
219
+ "grad_norm": 9.5625,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 1.0325,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15680684248039914,
226
+ "grad_norm": 9.1875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9825,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16393442622950818,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0758,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17106200997861726,
240
+ "grad_norm": 7.9375,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.076,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1781895937277263,
247
+ "grad_norm": 8.25,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0498,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18531717747683535,
254
+ "grad_norm": 7.9375,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9616,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.1924447612259444,
261
+ "grad_norm": 7.71875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 1.0782,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19957234497505347,
268
+ "grad_norm": 8.6875,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.9648,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.2066999287241625,
275
+ "grad_norm": 9.375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9675,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21382751247327156,
282
+ "grad_norm": 7.53125,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.9063,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2209550962223806,
289
+ "grad_norm": 6.96875,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.8631,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.22808267997148968,
296
+ "grad_norm": 7.28125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.9451,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23521026372059872,
303
+ "grad_norm": 8.9375,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 1.0807,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24233784746970777,
310
+ "grad_norm": 7.09375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9637,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.2494654312188168,
317
+ "grad_norm": 6.25,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9216,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25659301496792586,
324
+ "grad_norm": 6.34375,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.9588,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2637205987170349,
331
+ "grad_norm": 6.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8696,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.27084818246614395,
338
+ "grad_norm": 6.0625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.951,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.27797576621525305,
345
+ "grad_norm": 6.6875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.9381,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.2851033499643621,
352
+ "grad_norm": 7.3125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.9314,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29223093371347114,
359
+ "grad_norm": 5.84375,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9212,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2993585174625802,
366
+ "grad_norm": 6.90625,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.8897,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.30648610121168923,
373
+ "grad_norm": 5.46875,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.9813,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3136136849607983,
380
+ "grad_norm": 4.46875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.9023,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3207412687099073,
387
+ "grad_norm": 5.8125,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.882,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.32786885245901637,
394
+ "grad_norm": 6.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9943,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33499643620812547,
401
+ "grad_norm": 12.0,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8606,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.3421240199572345,
408
+ "grad_norm": 6.125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.9137,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34925160370634356,
415
+ "grad_norm": 6.25,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8677,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3563791874554526,
422
+ "grad_norm": 5.46875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.9161,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.36350677120456165,
429
+ "grad_norm": 6.125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.9176,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.3706343549536707,
436
+ "grad_norm": 6.34375,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.871,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.37776193870277974,
443
+ "grad_norm": 6.28125,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.9097,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.3848895224518888,
450
+ "grad_norm": 5.1875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8721,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.3920171062009979,
457
+ "grad_norm": 7.15625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.9282,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39914468995010693,
464
+ "grad_norm": 5.375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8213,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.406272273699216,
471
+ "grad_norm": 7.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8527,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.413399857448325,
478
+ "grad_norm": 5.75,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.944,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42052744119743407,
485
+ "grad_norm": 6.5625,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.9269,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.4276550249465431,
492
+ "grad_norm": 5.53125,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.8261,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43478260869565216,
499
+ "grad_norm": 6.4375,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 1.0049,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.4419101924447612,
506
+ "grad_norm": 6.65625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.9438,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44903777619387025,
513
+ "grad_norm": 5.9375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.9527,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.45616535994297935,
520
+ "grad_norm": 6.375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.8917,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.4632929436920884,
527
+ "grad_norm": 4.625,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8554,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47042052744119744,
534
+ "grad_norm": 5.59375,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.978,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.4775481111903065,
541
+ "grad_norm": 5.0625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.774,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48467569493941554,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8388,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4918032786885246,
555
+ "grad_norm": 4.8125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.8888,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4989308624376336,
562
+ "grad_norm": 6.03125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.7606,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5060584461867427,
569
+ "grad_norm": 5.1875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.9804,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5131860299358517,
576
+ "grad_norm": 6.5625,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.8832,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5203136136849608,
583
+ "grad_norm": 5.0,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.8385,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5274411974340698,
590
+ "grad_norm": 6.78125,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.9106,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5345687811831789,
597
+ "grad_norm": 6.3125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.807,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5416963649322879,
604
+ "grad_norm": 4.0,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7932,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5488239486813971,
611
+ "grad_norm": 5.53125,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.7533,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5559515324305061,
618
+ "grad_norm": 5.125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.8676,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5630791161796151,
625
+ "grad_norm": 5.0625,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8783,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5702066999287242,
632
+ "grad_norm": 5.5625,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.9352,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5773342836778332,
639
+ "grad_norm": 4.8125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8814,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5844618674269423,
646
+ "grad_norm": 5.96875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.8252,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5915894511760513,
653
+ "grad_norm": 4.96875,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8695,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5987170349251604,
660
+ "grad_norm": 5.21875,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8821,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6058446186742694,
667
+ "grad_norm": 6.375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.8968,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6129722024233785,
674
+ "grad_norm": 4.59375,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.9109,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6200997861724875,
681
+ "grad_norm": 6.625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.8701,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6272273699215966,
688
+ "grad_norm": 5.71875,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.8649,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6343549536707056,
695
+ "grad_norm": 5.25,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8578,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6414825374198146,
702
+ "grad_norm": 4.9375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.9067,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6486101211689237,
709
+ "grad_norm": 6.8125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7862,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6557377049180327,
716
+ "grad_norm": 5.875,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.8275,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6628652886671419,
723
+ "grad_norm": 5.6875,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.9155,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6699928724162509,
730
+ "grad_norm": 5.6875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.8526,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.67712045616536,
737
+ "grad_norm": 4.9375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8123,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.684248039914469,
744
+ "grad_norm": 5.4375,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7986,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6913756236635781,
751
+ "grad_norm": 5.6875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.9203,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6985032074126871,
758
+ "grad_norm": 5.78125,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7805,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7056307911617962,
765
+ "grad_norm": 5.4375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8683,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7127583749109052,
772
+ "grad_norm": 5.46875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8515,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7127583749109052,
779
+ "eval/nano_beir.arguana": 0.49904762134680597,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7127583749109052,
784
+ "eval/nano_beir.climate_fever": 0.2710368630075979,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7127583749109052,
789
+ "eval/nano_beir.dbpedia_entity": 0.32627448935866016,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7127583749109052,
794
+ "eval/nano_beir.fever": 0.6251056136949217,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7127583749109052,
799
+ "eval/nano_beir.fiqa": 0.4264627976119544,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7127583749109052,
804
+ "eval/nano_beir.hotpotqa": 0.67890984670666,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7127583749109052,
809
+ "eval/nano_beir.nfcorpus": 0.33575253271342886,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7127583749109052,
814
+ "eval/nano_beir.nq": 0.41529049449791683,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7127583749109052,
819
+ "eval/nano_beir.quora": 0.8400060722791932,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7127583749109052,
824
+ "eval/nano_beir.scidocs": 0.25236583848912914,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7127583749109052,
829
+ "eval/nano_beir.scifact": 0.7248783076751154,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7127583749109052,
834
+ "eval/nano_beir.webis_touche2020": 0.29804501841705877,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7127583749109052,
839
+ "eval/avg": 0.4744312913165369,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7198859586600143,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8998,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7270135424091233,
851
+ "grad_norm": 5.34375,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8763,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7341411261582323,
858
+ "grad_norm": 4.875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8851,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7412687099073414,
865
+ "grad_norm": 4.40625,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.8247,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7483962936564504,
872
+ "grad_norm": 6.5625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.742,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7555238774055595,
879
+ "grad_norm": 4.6875,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8707,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7626514611546685,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7698,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7697790449037776,
893
+ "grad_norm": 6.1875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.9318,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7769066286528866,
900
+ "grad_norm": 4.40625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.752,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7840342124019958,
907
+ "grad_norm": 6.0,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.9631,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7911617961511048,
914
+ "grad_norm": 4.71875,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.8623,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7982893799002139,
921
+ "grad_norm": 5.84375,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8677,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8054169636493229,
928
+ "grad_norm": 4.03125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.7795,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.812544547398432,
935
+ "grad_norm": 4.125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.8011,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.819672131147541,
942
+ "grad_norm": 4.875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7984,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.82679971489665,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.8049,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8339272986457591,
956
+ "grad_norm": 5.15625,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7839,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8410548823948681,
963
+ "grad_norm": 4.78125,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7809,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8481824661439772,
970
+ "grad_norm": 4.71875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.7688,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8553100498930862,
977
+ "grad_norm": 7.75,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.8638,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8624376336421953,
984
+ "grad_norm": 5.9375,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8409,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8695652173913043,
991
+ "grad_norm": 5.875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.916,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8766928011404134,
998
+ "grad_norm": 5.40625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.9114,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.8838203848895224,
1005
+ "grad_norm": 4.5,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.8744,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.8909479686386315,
1012
+ "grad_norm": 4.9375,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.8471,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8980755523877405,
1019
+ "grad_norm": 5.21875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8014,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9052031361368497,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.8587,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.9123307198859587,
1033
+ "grad_norm": 5.9375,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8003,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.9194583036350678,
1040
+ "grad_norm": 6.09375,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7404,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.9265858873841768,
1047
+ "grad_norm": 5.5,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.722,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9337134711332858,
1054
+ "grad_norm": 5.21875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7005,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9408410548823949,
1061
+ "grad_norm": 6.78125,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.7987,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.9479686386315039,
1068
+ "grad_norm": 5.78125,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7872,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.955096222380613,
1075
+ "grad_norm": 4.875,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.8632,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.962223806129722,
1082
+ "grad_norm": 6.96875,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.8165,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9693513898788311,
1089
+ "grad_norm": 5.0625,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7802,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9764789736279401,
1096
+ "grad_norm": 5.34375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.8668,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9836065573770492,
1103
+ "grad_norm": 5.84375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.7785,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9907341411261582,
1110
+ "grad_norm": 6.25,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8952,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9978617248752673,
1117
+ "grad_norm": 6.34375,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.826,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.0049893086243764,
1124
+ "grad_norm": 6.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.9168,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0121168923734853,
1131
+ "grad_norm": 5.625,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8506,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0192444761225945,
1138
+ "grad_norm": 4.875,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.9257,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0263720598717034,
1145
+ "grad_norm": 5.96875,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.8302,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0334996436208126,
1152
+ "grad_norm": 5.375,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.7898,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0406272273699215,
1159
+ "grad_norm": 5.75,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.7682,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0477548111190307,
1166
+ "grad_norm": 4.9375,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.8378,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0548823948681396,
1173
+ "grad_norm": 4.15625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.7113,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0620099786172488,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.7187,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0691375623663577,
1187
+ "grad_norm": 5.21875,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.7843,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0762651461154669,
1194
+ "grad_norm": 5.28125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.7253,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0833927298645758,
1201
+ "grad_norm": 4.78125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.8077,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.090520313613685,
1208
+ "grad_norm": 5.8125,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.8465,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0976478973627941,
1215
+ "grad_norm": 4.875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.7467,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.104775481111903,
1222
+ "grad_norm": 5.0,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.7919,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1119030648610122,
1229
+ "grad_norm": 5.78125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7399,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1190306486101211,
1236
+ "grad_norm": 5.28125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.8305,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1261582323592303,
1243
+ "grad_norm": 5.09375,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.8986,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1332858161083392,
1250
+ "grad_norm": 5.65625,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8041,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1404133998574484,
1257
+ "grad_norm": 5.5625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8434,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1475409836065573,
1264
+ "grad_norm": 5.9375,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7681,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1546685673556665,
1271
+ "grad_norm": 4.71875,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.6835,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1617961511047754,
1278
+ "grad_norm": 4.8125,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.8917,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1689237348538846,
1285
+ "grad_norm": 5.25,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8296,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1760513186029935,
1292
+ "grad_norm": 4.90625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8147,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1831789023521027,
1299
+ "grad_norm": 6.90625,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.7656,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1903064861012118,
1306
+ "grad_norm": 5.28125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8931,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1974340698503207,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.7844,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2045616535994297,
1320
+ "grad_norm": 5.53125,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7334,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2116892373485388,
1327
+ "grad_norm": 5.28125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7509,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.218816821097648,
1334
+ "grad_norm": 5.5,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8287,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.225944404846757,
1341
+ "grad_norm": 5.4375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.8383,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.233071988595866,
1348
+ "grad_norm": 6.6875,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7443,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.240199572344975,
1355
+ "grad_norm": 6.0625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7868,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2473271560940842,
1362
+ "grad_norm": 5.09375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.7662,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2544547398431931,
1369
+ "grad_norm": 4.90625,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7558,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2615823235923023,
1376
+ "grad_norm": 5.34375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6894,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2687099073414112,
1383
+ "grad_norm": 5.15625,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7782,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2758374910905204,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.756,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2829650748396293,
1397
+ "grad_norm": 4.125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.763,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.2900926585887385,
1404
+ "grad_norm": 10.5,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7255,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.2972202423378474,
1411
+ "grad_norm": 4.65625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7615,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3043478260869565,
1418
+ "grad_norm": 5.15625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8092,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3114754098360657,
1425
+ "grad_norm": 4.71875,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7503,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3186029935851746,
1432
+ "grad_norm": 4.5,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.8121,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.3257305773342836,
1439
+ "grad_norm": 4.78125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.8773,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.3328581610833927,
1446
+ "grad_norm": 5.6875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.8202,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.3399857448325019,
1453
+ "grad_norm": 6.15625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.8359,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3471133285816108,
1460
+ "grad_norm": 5.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.9009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.3542409123307197,
1467
+ "grad_norm": 4.96875,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.7515,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.361368496079829,
1474
+ "grad_norm": 4.96875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.8321,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.368496079828938,
1481
+ "grad_norm": 6.875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.8211,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.375623663578047,
1488
+ "grad_norm": 3.921875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7751,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.3827512473271562,
1495
+ "grad_norm": 5.53125,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8174,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.389878831076265,
1502
+ "grad_norm": 5.28125,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7058,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3970064148253742,
1509
+ "grad_norm": 6.34375,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7263,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.4041339985744832,
1516
+ "grad_norm": 5.5625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.7503,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.4112615823235923,
1523
+ "grad_norm": 5.34375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.87,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4183891660727013,
1530
+ "grad_norm": 4.84375,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7866,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4255167498218104,
1537
+ "grad_norm": 5.0625,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.787,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.4255167498218104,
1544
+ "eval/nano_beir.arguana": 0.45921477926341064,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.4255167498218104,
1549
+ "eval/nano_beir.climate_fever": 0.26026043643827973,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.4255167498218104,
1554
+ "eval/nano_beir.dbpedia_entity": 0.32492910818341564,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.4255167498218104,
1559
+ "eval/nano_beir.fever": 0.6243268043774819,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.4255167498218104,
1564
+ "eval/nano_beir.fiqa": 0.406434177003067,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.4255167498218104,
1569
+ "eval/nano_beir.hotpotqa": 0.6761769995795371,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.4255167498218104,
1574
+ "eval/nano_beir.nfcorpus": 0.3490707285646159,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.4255167498218104,
1579
+ "eval/nano_beir.nq": 0.40743688778055903,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.4255167498218104,
1584
+ "eval/nano_beir.quora": 0.8329214304594091,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.4255167498218104,
1589
+ "eval/nano_beir.scidocs": 0.2444219693805352,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.4255167498218104,
1594
+ "eval/nano_beir.scifact": 0.7389917242185631,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.4255167498218104,
1599
+ "eval/nano_beir.webis_touche2020": 0.2912623613111957,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.4255167498218104,
1604
+ "eval/avg": 0.4679539505466725,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.4326443335709196,
1609
+ "grad_norm": 5.53125,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.873,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.4397719173200285,
1616
+ "grad_norm": 6.34375,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.8482,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.4468995010691375,
1623
+ "grad_norm": 4.875,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.8179,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.4540270848182466,
1630
+ "grad_norm": 4.40625,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.7916,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.4611546685673558,
1637
+ "grad_norm": 4.84375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.7528,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.4682822523164647,
1644
+ "grad_norm": 5.75,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.7749,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.4754098360655736,
1651
+ "grad_norm": 5.625,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.8334,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.4825374198146828,
1658
+ "grad_norm": 7.03125,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.7086,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.489665003563792,
1665
+ "grad_norm": 5.25,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.7891,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.4967925873129009,
1672
+ "grad_norm": 5.0625,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.7186,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.5039201710620098,
1679
+ "grad_norm": 5.28125,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.778,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.511047754811119,
1686
+ "grad_norm": 4.875,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.7799,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.5181753385602281,
1693
+ "grad_norm": 6.09375,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.7962,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.5253029223093373,
1700
+ "grad_norm": 5.84375,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.8263,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.5324305060584462,
1707
+ "grad_norm": 4.78125,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.7541,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.5395580898075552,
1714
+ "grad_norm": 5.1875,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.7617,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.5466856735566643,
1721
+ "grad_norm": 6.28125,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.8576,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.5538132573057735,
1728
+ "grad_norm": 5.6875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.7801,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.5609408410548824,
1735
+ "grad_norm": 5.96875,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.7256,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.5680684248039913,
1742
+ "grad_norm": 4.03125,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.7574,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.5751960085531005,
1749
+ "grad_norm": 6.65625,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.7539,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.5823235923022096,
1756
+ "grad_norm": 6.25,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.7865,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.5894511760513186,
1763
+ "grad_norm": 4.90625,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.781,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.5965787598004275,
1770
+ "grad_norm": 6.5625,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.9054,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.6037063435495367,
1777
+ "grad_norm": 4.0625,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.7958,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.6108339272986458,
1784
+ "grad_norm": 5.8125,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.7319,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.6179615110477548,
1791
+ "grad_norm": 5.5625,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.7538,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.6250890947968637,
1798
+ "grad_norm": 4.625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.7781,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.6322166785459729,
1805
+ "grad_norm": 6.21875,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.7636,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.639344262295082,
1812
+ "grad_norm": 5.65625,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.7041,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.6464718460441912,
1819
+ "grad_norm": 4.53125,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.8014,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.6535994297933,
1826
+ "grad_norm": 4.8125,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.7043,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.660727013542409,
1833
+ "grad_norm": 4.78125,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.7739,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.6678545972915182,
1840
+ "grad_norm": 5.0625,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.7807,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.6749821810406273,
1847
+ "grad_norm": 4.84375,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.8122,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.6821097647897363,
1854
+ "grad_norm": 6.25,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.8824,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.6892373485388452,
1861
+ "grad_norm": 4.46875,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.6925,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.6963649322879544,
1868
+ "grad_norm": 5.59375,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.7558,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.7034925160370635,
1875
+ "grad_norm": 5.09375,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.7045,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.7106200997861725,
1882
+ "grad_norm": 5.0625,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.7654,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.7177476835352814,
1889
+ "grad_norm": 4.90625,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.7696,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.7248752672843906,
1896
+ "grad_norm": 5.65625,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.8205,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.7320028510334997,
1903
+ "grad_norm": 6.1875,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.7752,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.7391304347826086,
1910
+ "grad_norm": 4.4375,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.79,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.7462580185317176,
1917
+ "grad_norm": 4.75,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.9032,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.7533856022808267,
1924
+ "grad_norm": 5.09375,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.7631,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.760513186029936,
1931
+ "grad_norm": 4.9375,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.7612,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.767640769779045,
1938
+ "grad_norm": 5.125,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.7827,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.774768353528154,
1945
+ "grad_norm": 4.8125,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.7827,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.781895937277263,
1952
+ "grad_norm": 4.6875,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.7646,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.789023521026372,
1959
+ "grad_norm": 7.3125,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.8215,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.7961511047754812,
1966
+ "grad_norm": 4.84375,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.7416,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.8032786885245902,
1973
+ "grad_norm": 4.3125,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.7577,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.810406272273699,
1980
+ "grad_norm": 4.34375,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.6839,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 1.8175338560228083,
1987
+ "grad_norm": 4.46875,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.719,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 1.8246614397719174,
1994
+ "grad_norm": 5.875,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.7487,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 1.8317890235210263,
2001
+ "grad_norm": 5.40625,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.7946,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 1.8389166072701353,
2008
+ "grad_norm": 5.59375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.7822,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 1.8460441910192444,
2015
+ "grad_norm": 6.59375,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.7763,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 1.8531717747683536,
2022
+ "grad_norm": 5.90625,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.8152,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 1.8602993585174625,
2029
+ "grad_norm": 5.53125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.8451,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 1.8674269422665715,
2036
+ "grad_norm": 4.6875,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.7819,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 1.8745545260156806,
2043
+ "grad_norm": 5.3125,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.794,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 1.8816821097647898,
2050
+ "grad_norm": 4.8125,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.7435,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 1.888809693513899,
2057
+ "grad_norm": 6.09375,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.8852,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 1.8959372772630079,
2064
+ "grad_norm": 5.46875,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.7538,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 1.9030648610121168,
2071
+ "grad_norm": 5.3125,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.7795,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 1.910192444761226,
2078
+ "grad_norm": 5.65625,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.7887,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 1.9173200285103351,
2085
+ "grad_norm": 6.09375,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.6838,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 1.924447612259444,
2092
+ "grad_norm": 6.4375,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.7495,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 1.931575196008553,
2099
+ "grad_norm": 5.34375,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.7499,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 1.9387027797576621,
2106
+ "grad_norm": 5.21875,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.8243,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 1.9458303635067713,
2113
+ "grad_norm": 5.34375,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.8128,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 1.9529579472558802,
2120
+ "grad_norm": 5.71875,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.7626,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 1.9600855310049892,
2127
+ "grad_norm": 4.96875,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.7724,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 1.9672131147540983,
2134
+ "grad_norm": 5.59375,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.7741,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 1.9743406985032075,
2141
+ "grad_norm": 6.3125,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.7482,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 1.9814682822523164,
2148
+ "grad_norm": 5.6875,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.7981,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 1.9885958660014256,
2155
+ "grad_norm": 4.875,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.7901,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 1.9957234497505345,
2162
+ "grad_norm": 5.96875,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.7576,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.0028510334996437,
2169
+ "grad_norm": 5.4375,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.7789,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.009978617248753,
2176
+ "grad_norm": 4.96875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.8136,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.0171062009978615,
2183
+ "grad_norm": 5.1875,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.813,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.0242337847469707,
2190
+ "grad_norm": 5.96875,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.7948,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.03136136849608,
2197
+ "grad_norm": 5.09375,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6773,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.038488952245189,
2204
+ "grad_norm": 5.03125,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.6645,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.0456165359942977,
2211
+ "grad_norm": 5.34375,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.7572,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.052744119743407,
2218
+ "grad_norm": 4.59375,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.7588,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.059871703492516,
2225
+ "grad_norm": 5.625,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.7294,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.066999287241625,
2232
+ "grad_norm": 5.09375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.7621,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.0741268709907343,
2239
+ "grad_norm": 6.65625,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.8967,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.081254454739843,
2246
+ "grad_norm": 5.3125,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.84,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.088382038488952,
2253
+ "grad_norm": 4.375,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.7584,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.0955096222380614,
2260
+ "grad_norm": 5.21875,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.8475,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.1026372059871705,
2267
+ "grad_norm": 6.40625,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.7067,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.1097647897362792,
2274
+ "grad_norm": 6.1875,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.8023,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.1168923734853884,
2281
+ "grad_norm": 5.03125,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.7074,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.1240199572344975,
2288
+ "grad_norm": 7.5,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.7498,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.1311475409836067,
2295
+ "grad_norm": 7.0625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.7249,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.1382751247327154,
2302
+ "grad_norm": 5.53125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.7907,
2305
+ "step": 3000
2306
+ }
2307
+ ],
2308
+ "logging_steps": 10,
2309
+ "max_steps": 5000,
2310
+ "num_input_tokens_seen": 0,
2311
+ "num_train_epochs": 4,
2312
+ "save_steps": 1000,
2313
+ "stateful_callbacks": {
2314
+ "TrainerControl": {
2315
+ "args": {
2316
+ "should_epoch_stop": false,
2317
+ "should_evaluate": false,
2318
+ "should_log": false,
2319
+ "should_save": true,
2320
+ "should_training_stop": false
2321
+ },
2322
+ "attributes": {}
2323
+ }
2324
+ },
2325
+ "total_flos": 0.0,
2326
+ "train_batch_size": 16,
2327
+ "trial_name": null,
2328
+ "trial_params": null
2329
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-3000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1afec67e7d18e7478623434fcf8c1bcc7719c01c4f7874089687d7138953afff
3
+ size 6097
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b72e747ec081562370d86737cb071fdb265789168f91e384f6b53c8d0ab05200
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bf67c8d687a414eecf2c66a278f63643cc7e1dbaa26d63ca88af5b5188693ff
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0e770c61e3acef7bd144efee65a450ed487b244ed293a8dd801394ffcae775
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21354c9980d6ba769d430eb9962ffba457eb04a66cb90e33b0934d3157fae7cf
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d574a8567b1636a399633d2b20273fdb3be4b888e37f0d331cd695b973f10463
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cca2b80df90b78b3ba9d9d2951c5b7aa80fc0a44b4512b495de6de40396fcb6e
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9218a8f161ba45a13a24f24cd020f32b0abec4bcc53a2764e48b480568dd2d13
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/trainer_state.json ADDED
@@ -0,0 +1,3094 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.851033499643621,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 4000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007127583749109052,
79
+ "grad_norm": 11.9375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.5314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014255167498218105,
86
+ "grad_norm": 11.125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.4333,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021382751247327157,
93
+ "grad_norm": 22.75,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.6002,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.02851033499643621,
100
+ "grad_norm": 13.3125,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.5692,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03563791874554526,
107
+ "grad_norm": 14.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.2221,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042765502494654314,
114
+ "grad_norm": 12.4375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.3981,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04989308624376337,
121
+ "grad_norm": 11.375,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.3844,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05702066999287242,
128
+ "grad_norm": 12.375,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.6756,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06414825374198146,
135
+ "grad_norm": 9.8125,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3874,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07127583749109052,
142
+ "grad_norm": 11.5,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.3705,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07840342124019957,
149
+ "grad_norm": 10.0,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3897,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08553100498930863,
156
+ "grad_norm": 11.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.4046,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09265858873841767,
163
+ "grad_norm": 8.25,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1084,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09978617248752673,
170
+ "grad_norm": 7.96875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.1872,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10691375623663578,
177
+ "grad_norm": 10.8125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.4332,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11404133998574484,
184
+ "grad_norm": 12.8125,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2686,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12116892373485388,
191
+ "grad_norm": 12.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.9751,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12829650748396293,
198
+ "grad_norm": 9.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1097,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.13542409123307197,
205
+ "grad_norm": 10.6875,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 1.0446,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14255167498218105,
212
+ "grad_norm": 8.625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.2052,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.1496792587312901,
219
+ "grad_norm": 9.5625,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 1.0325,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15680684248039914,
226
+ "grad_norm": 9.1875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9825,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16393442622950818,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0758,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17106200997861726,
240
+ "grad_norm": 7.9375,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.076,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1781895937277263,
247
+ "grad_norm": 8.25,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0498,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18531717747683535,
254
+ "grad_norm": 7.9375,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9616,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.1924447612259444,
261
+ "grad_norm": 7.71875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 1.0782,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19957234497505347,
268
+ "grad_norm": 8.6875,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.9648,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.2066999287241625,
275
+ "grad_norm": 9.375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9675,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21382751247327156,
282
+ "grad_norm": 7.53125,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.9063,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2209550962223806,
289
+ "grad_norm": 6.96875,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.8631,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.22808267997148968,
296
+ "grad_norm": 7.28125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.9451,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23521026372059872,
303
+ "grad_norm": 8.9375,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 1.0807,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24233784746970777,
310
+ "grad_norm": 7.09375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9637,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.2494654312188168,
317
+ "grad_norm": 6.25,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9216,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25659301496792586,
324
+ "grad_norm": 6.34375,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.9588,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2637205987170349,
331
+ "grad_norm": 6.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8696,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.27084818246614395,
338
+ "grad_norm": 6.0625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.951,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.27797576621525305,
345
+ "grad_norm": 6.6875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.9381,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.2851033499643621,
352
+ "grad_norm": 7.3125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.9314,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29223093371347114,
359
+ "grad_norm": 5.84375,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9212,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2993585174625802,
366
+ "grad_norm": 6.90625,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.8897,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.30648610121168923,
373
+ "grad_norm": 5.46875,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.9813,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3136136849607983,
380
+ "grad_norm": 4.46875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.9023,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3207412687099073,
387
+ "grad_norm": 5.8125,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.882,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.32786885245901637,
394
+ "grad_norm": 6.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9943,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33499643620812547,
401
+ "grad_norm": 12.0,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8606,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.3421240199572345,
408
+ "grad_norm": 6.125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.9137,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34925160370634356,
415
+ "grad_norm": 6.25,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8677,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3563791874554526,
422
+ "grad_norm": 5.46875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.9161,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.36350677120456165,
429
+ "grad_norm": 6.125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.9176,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.3706343549536707,
436
+ "grad_norm": 6.34375,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.871,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.37776193870277974,
443
+ "grad_norm": 6.28125,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.9097,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.3848895224518888,
450
+ "grad_norm": 5.1875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8721,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.3920171062009979,
457
+ "grad_norm": 7.15625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.9282,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39914468995010693,
464
+ "grad_norm": 5.375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8213,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.406272273699216,
471
+ "grad_norm": 7.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8527,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.413399857448325,
478
+ "grad_norm": 5.75,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.944,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42052744119743407,
485
+ "grad_norm": 6.5625,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.9269,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.4276550249465431,
492
+ "grad_norm": 5.53125,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.8261,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43478260869565216,
499
+ "grad_norm": 6.4375,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 1.0049,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.4419101924447612,
506
+ "grad_norm": 6.65625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.9438,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44903777619387025,
513
+ "grad_norm": 5.9375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.9527,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.45616535994297935,
520
+ "grad_norm": 6.375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.8917,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.4632929436920884,
527
+ "grad_norm": 4.625,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8554,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47042052744119744,
534
+ "grad_norm": 5.59375,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.978,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.4775481111903065,
541
+ "grad_norm": 5.0625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.774,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48467569493941554,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8388,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4918032786885246,
555
+ "grad_norm": 4.8125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.8888,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4989308624376336,
562
+ "grad_norm": 6.03125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.7606,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5060584461867427,
569
+ "grad_norm": 5.1875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.9804,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5131860299358517,
576
+ "grad_norm": 6.5625,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.8832,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5203136136849608,
583
+ "grad_norm": 5.0,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.8385,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5274411974340698,
590
+ "grad_norm": 6.78125,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.9106,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5345687811831789,
597
+ "grad_norm": 6.3125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.807,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5416963649322879,
604
+ "grad_norm": 4.0,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7932,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5488239486813971,
611
+ "grad_norm": 5.53125,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.7533,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5559515324305061,
618
+ "grad_norm": 5.125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.8676,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5630791161796151,
625
+ "grad_norm": 5.0625,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8783,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5702066999287242,
632
+ "grad_norm": 5.5625,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.9352,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5773342836778332,
639
+ "grad_norm": 4.8125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8814,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5844618674269423,
646
+ "grad_norm": 5.96875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.8252,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5915894511760513,
653
+ "grad_norm": 4.96875,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8695,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5987170349251604,
660
+ "grad_norm": 5.21875,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8821,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6058446186742694,
667
+ "grad_norm": 6.375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.8968,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6129722024233785,
674
+ "grad_norm": 4.59375,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.9109,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6200997861724875,
681
+ "grad_norm": 6.625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.8701,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6272273699215966,
688
+ "grad_norm": 5.71875,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.8649,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6343549536707056,
695
+ "grad_norm": 5.25,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8578,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6414825374198146,
702
+ "grad_norm": 4.9375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.9067,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6486101211689237,
709
+ "grad_norm": 6.8125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7862,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6557377049180327,
716
+ "grad_norm": 5.875,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.8275,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6628652886671419,
723
+ "grad_norm": 5.6875,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.9155,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6699928724162509,
730
+ "grad_norm": 5.6875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.8526,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.67712045616536,
737
+ "grad_norm": 4.9375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8123,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.684248039914469,
744
+ "grad_norm": 5.4375,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7986,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6913756236635781,
751
+ "grad_norm": 5.6875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.9203,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6985032074126871,
758
+ "grad_norm": 5.78125,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7805,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7056307911617962,
765
+ "grad_norm": 5.4375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8683,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7127583749109052,
772
+ "grad_norm": 5.46875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8515,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7127583749109052,
779
+ "eval/nano_beir.arguana": 0.49904762134680597,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7127583749109052,
784
+ "eval/nano_beir.climate_fever": 0.2710368630075979,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7127583749109052,
789
+ "eval/nano_beir.dbpedia_entity": 0.32627448935866016,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7127583749109052,
794
+ "eval/nano_beir.fever": 0.6251056136949217,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7127583749109052,
799
+ "eval/nano_beir.fiqa": 0.4264627976119544,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7127583749109052,
804
+ "eval/nano_beir.hotpotqa": 0.67890984670666,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7127583749109052,
809
+ "eval/nano_beir.nfcorpus": 0.33575253271342886,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7127583749109052,
814
+ "eval/nano_beir.nq": 0.41529049449791683,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7127583749109052,
819
+ "eval/nano_beir.quora": 0.8400060722791932,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7127583749109052,
824
+ "eval/nano_beir.scidocs": 0.25236583848912914,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7127583749109052,
829
+ "eval/nano_beir.scifact": 0.7248783076751154,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7127583749109052,
834
+ "eval/nano_beir.webis_touche2020": 0.29804501841705877,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7127583749109052,
839
+ "eval/avg": 0.4744312913165369,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7198859586600143,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8998,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7270135424091233,
851
+ "grad_norm": 5.34375,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8763,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7341411261582323,
858
+ "grad_norm": 4.875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8851,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7412687099073414,
865
+ "grad_norm": 4.40625,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.8247,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7483962936564504,
872
+ "grad_norm": 6.5625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.742,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7555238774055595,
879
+ "grad_norm": 4.6875,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8707,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7626514611546685,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7698,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7697790449037776,
893
+ "grad_norm": 6.1875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.9318,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7769066286528866,
900
+ "grad_norm": 4.40625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.752,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7840342124019958,
907
+ "grad_norm": 6.0,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.9631,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7911617961511048,
914
+ "grad_norm": 4.71875,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.8623,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7982893799002139,
921
+ "grad_norm": 5.84375,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8677,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8054169636493229,
928
+ "grad_norm": 4.03125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.7795,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.812544547398432,
935
+ "grad_norm": 4.125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.8011,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.819672131147541,
942
+ "grad_norm": 4.875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7984,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.82679971489665,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.8049,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8339272986457591,
956
+ "grad_norm": 5.15625,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7839,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8410548823948681,
963
+ "grad_norm": 4.78125,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7809,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8481824661439772,
970
+ "grad_norm": 4.71875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.7688,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8553100498930862,
977
+ "grad_norm": 7.75,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.8638,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8624376336421953,
984
+ "grad_norm": 5.9375,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8409,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8695652173913043,
991
+ "grad_norm": 5.875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.916,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8766928011404134,
998
+ "grad_norm": 5.40625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.9114,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.8838203848895224,
1005
+ "grad_norm": 4.5,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.8744,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.8909479686386315,
1012
+ "grad_norm": 4.9375,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.8471,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8980755523877405,
1019
+ "grad_norm": 5.21875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8014,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9052031361368497,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.8587,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.9123307198859587,
1033
+ "grad_norm": 5.9375,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8003,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.9194583036350678,
1040
+ "grad_norm": 6.09375,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7404,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.9265858873841768,
1047
+ "grad_norm": 5.5,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.722,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9337134711332858,
1054
+ "grad_norm": 5.21875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7005,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9408410548823949,
1061
+ "grad_norm": 6.78125,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.7987,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.9479686386315039,
1068
+ "grad_norm": 5.78125,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7872,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.955096222380613,
1075
+ "grad_norm": 4.875,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.8632,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.962223806129722,
1082
+ "grad_norm": 6.96875,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.8165,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9693513898788311,
1089
+ "grad_norm": 5.0625,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7802,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9764789736279401,
1096
+ "grad_norm": 5.34375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.8668,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9836065573770492,
1103
+ "grad_norm": 5.84375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.7785,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9907341411261582,
1110
+ "grad_norm": 6.25,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8952,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9978617248752673,
1117
+ "grad_norm": 6.34375,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.826,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.0049893086243764,
1124
+ "grad_norm": 6.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.9168,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0121168923734853,
1131
+ "grad_norm": 5.625,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8506,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0192444761225945,
1138
+ "grad_norm": 4.875,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.9257,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0263720598717034,
1145
+ "grad_norm": 5.96875,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.8302,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0334996436208126,
1152
+ "grad_norm": 5.375,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.7898,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0406272273699215,
1159
+ "grad_norm": 5.75,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.7682,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0477548111190307,
1166
+ "grad_norm": 4.9375,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.8378,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0548823948681396,
1173
+ "grad_norm": 4.15625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.7113,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0620099786172488,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.7187,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0691375623663577,
1187
+ "grad_norm": 5.21875,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.7843,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0762651461154669,
1194
+ "grad_norm": 5.28125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.7253,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0833927298645758,
1201
+ "grad_norm": 4.78125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.8077,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.090520313613685,
1208
+ "grad_norm": 5.8125,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.8465,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0976478973627941,
1215
+ "grad_norm": 4.875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.7467,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.104775481111903,
1222
+ "grad_norm": 5.0,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.7919,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1119030648610122,
1229
+ "grad_norm": 5.78125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7399,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1190306486101211,
1236
+ "grad_norm": 5.28125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.8305,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1261582323592303,
1243
+ "grad_norm": 5.09375,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.8986,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1332858161083392,
1250
+ "grad_norm": 5.65625,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8041,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1404133998574484,
1257
+ "grad_norm": 5.5625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8434,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1475409836065573,
1264
+ "grad_norm": 5.9375,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7681,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1546685673556665,
1271
+ "grad_norm": 4.71875,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.6835,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1617961511047754,
1278
+ "grad_norm": 4.8125,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.8917,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1689237348538846,
1285
+ "grad_norm": 5.25,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8296,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1760513186029935,
1292
+ "grad_norm": 4.90625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8147,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1831789023521027,
1299
+ "grad_norm": 6.90625,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.7656,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1903064861012118,
1306
+ "grad_norm": 5.28125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8931,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1974340698503207,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.7844,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2045616535994297,
1320
+ "grad_norm": 5.53125,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7334,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2116892373485388,
1327
+ "grad_norm": 5.28125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7509,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.218816821097648,
1334
+ "grad_norm": 5.5,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8287,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.225944404846757,
1341
+ "grad_norm": 5.4375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.8383,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.233071988595866,
1348
+ "grad_norm": 6.6875,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7443,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.240199572344975,
1355
+ "grad_norm": 6.0625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7868,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2473271560940842,
1362
+ "grad_norm": 5.09375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.7662,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2544547398431931,
1369
+ "grad_norm": 4.90625,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7558,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2615823235923023,
1376
+ "grad_norm": 5.34375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6894,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2687099073414112,
1383
+ "grad_norm": 5.15625,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7782,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2758374910905204,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.756,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2829650748396293,
1397
+ "grad_norm": 4.125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.763,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.2900926585887385,
1404
+ "grad_norm": 10.5,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7255,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.2972202423378474,
1411
+ "grad_norm": 4.65625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7615,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3043478260869565,
1418
+ "grad_norm": 5.15625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8092,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3114754098360657,
1425
+ "grad_norm": 4.71875,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7503,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3186029935851746,
1432
+ "grad_norm": 4.5,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.8121,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.3257305773342836,
1439
+ "grad_norm": 4.78125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.8773,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.3328581610833927,
1446
+ "grad_norm": 5.6875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.8202,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.3399857448325019,
1453
+ "grad_norm": 6.15625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.8359,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3471133285816108,
1460
+ "grad_norm": 5.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.9009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.3542409123307197,
1467
+ "grad_norm": 4.96875,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.7515,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.361368496079829,
1474
+ "grad_norm": 4.96875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.8321,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.368496079828938,
1481
+ "grad_norm": 6.875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.8211,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.375623663578047,
1488
+ "grad_norm": 3.921875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7751,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.3827512473271562,
1495
+ "grad_norm": 5.53125,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8174,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.389878831076265,
1502
+ "grad_norm": 5.28125,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7058,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3970064148253742,
1509
+ "grad_norm": 6.34375,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7263,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.4041339985744832,
1516
+ "grad_norm": 5.5625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.7503,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.4112615823235923,
1523
+ "grad_norm": 5.34375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.87,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4183891660727013,
1530
+ "grad_norm": 4.84375,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7866,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4255167498218104,
1537
+ "grad_norm": 5.0625,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.787,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.4255167498218104,
1544
+ "eval/nano_beir.arguana": 0.45921477926341064,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.4255167498218104,
1549
+ "eval/nano_beir.climate_fever": 0.26026043643827973,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.4255167498218104,
1554
+ "eval/nano_beir.dbpedia_entity": 0.32492910818341564,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.4255167498218104,
1559
+ "eval/nano_beir.fever": 0.6243268043774819,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.4255167498218104,
1564
+ "eval/nano_beir.fiqa": 0.406434177003067,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.4255167498218104,
1569
+ "eval/nano_beir.hotpotqa": 0.6761769995795371,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.4255167498218104,
1574
+ "eval/nano_beir.nfcorpus": 0.3490707285646159,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.4255167498218104,
1579
+ "eval/nano_beir.nq": 0.40743688778055903,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.4255167498218104,
1584
+ "eval/nano_beir.quora": 0.8329214304594091,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.4255167498218104,
1589
+ "eval/nano_beir.scidocs": 0.2444219693805352,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.4255167498218104,
1594
+ "eval/nano_beir.scifact": 0.7389917242185631,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.4255167498218104,
1599
+ "eval/nano_beir.webis_touche2020": 0.2912623613111957,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.4255167498218104,
1604
+ "eval/avg": 0.4679539505466725,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.4326443335709196,
1609
+ "grad_norm": 5.53125,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.873,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.4397719173200285,
1616
+ "grad_norm": 6.34375,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.8482,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.4468995010691375,
1623
+ "grad_norm": 4.875,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.8179,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.4540270848182466,
1630
+ "grad_norm": 4.40625,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.7916,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.4611546685673558,
1637
+ "grad_norm": 4.84375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.7528,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.4682822523164647,
1644
+ "grad_norm": 5.75,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.7749,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.4754098360655736,
1651
+ "grad_norm": 5.625,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.8334,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.4825374198146828,
1658
+ "grad_norm": 7.03125,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.7086,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.489665003563792,
1665
+ "grad_norm": 5.25,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.7891,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.4967925873129009,
1672
+ "grad_norm": 5.0625,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.7186,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.5039201710620098,
1679
+ "grad_norm": 5.28125,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.778,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.511047754811119,
1686
+ "grad_norm": 4.875,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.7799,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.5181753385602281,
1693
+ "grad_norm": 6.09375,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.7962,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.5253029223093373,
1700
+ "grad_norm": 5.84375,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.8263,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.5324305060584462,
1707
+ "grad_norm": 4.78125,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.7541,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.5395580898075552,
1714
+ "grad_norm": 5.1875,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.7617,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.5466856735566643,
1721
+ "grad_norm": 6.28125,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.8576,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.5538132573057735,
1728
+ "grad_norm": 5.6875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.7801,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.5609408410548824,
1735
+ "grad_norm": 5.96875,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.7256,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.5680684248039913,
1742
+ "grad_norm": 4.03125,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.7574,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.5751960085531005,
1749
+ "grad_norm": 6.65625,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.7539,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.5823235923022096,
1756
+ "grad_norm": 6.25,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.7865,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.5894511760513186,
1763
+ "grad_norm": 4.90625,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.781,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.5965787598004275,
1770
+ "grad_norm": 6.5625,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.9054,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.6037063435495367,
1777
+ "grad_norm": 4.0625,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.7958,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.6108339272986458,
1784
+ "grad_norm": 5.8125,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.7319,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.6179615110477548,
1791
+ "grad_norm": 5.5625,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.7538,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.6250890947968637,
1798
+ "grad_norm": 4.625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.7781,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.6322166785459729,
1805
+ "grad_norm": 6.21875,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.7636,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.639344262295082,
1812
+ "grad_norm": 5.65625,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.7041,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.6464718460441912,
1819
+ "grad_norm": 4.53125,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.8014,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.6535994297933,
1826
+ "grad_norm": 4.8125,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.7043,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.660727013542409,
1833
+ "grad_norm": 4.78125,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.7739,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.6678545972915182,
1840
+ "grad_norm": 5.0625,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.7807,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.6749821810406273,
1847
+ "grad_norm": 4.84375,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.8122,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.6821097647897363,
1854
+ "grad_norm": 6.25,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.8824,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.6892373485388452,
1861
+ "grad_norm": 4.46875,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.6925,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.6963649322879544,
1868
+ "grad_norm": 5.59375,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.7558,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.7034925160370635,
1875
+ "grad_norm": 5.09375,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.7045,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.7106200997861725,
1882
+ "grad_norm": 5.0625,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.7654,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.7177476835352814,
1889
+ "grad_norm": 4.90625,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.7696,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.7248752672843906,
1896
+ "grad_norm": 5.65625,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.8205,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.7320028510334997,
1903
+ "grad_norm": 6.1875,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.7752,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.7391304347826086,
1910
+ "grad_norm": 4.4375,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.79,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.7462580185317176,
1917
+ "grad_norm": 4.75,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.9032,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.7533856022808267,
1924
+ "grad_norm": 5.09375,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.7631,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.760513186029936,
1931
+ "grad_norm": 4.9375,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.7612,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.767640769779045,
1938
+ "grad_norm": 5.125,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.7827,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.774768353528154,
1945
+ "grad_norm": 4.8125,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.7827,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.781895937277263,
1952
+ "grad_norm": 4.6875,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.7646,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.789023521026372,
1959
+ "grad_norm": 7.3125,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.8215,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.7961511047754812,
1966
+ "grad_norm": 4.84375,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.7416,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.8032786885245902,
1973
+ "grad_norm": 4.3125,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.7577,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.810406272273699,
1980
+ "grad_norm": 4.34375,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.6839,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 1.8175338560228083,
1987
+ "grad_norm": 4.46875,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.719,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 1.8246614397719174,
1994
+ "grad_norm": 5.875,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.7487,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 1.8317890235210263,
2001
+ "grad_norm": 5.40625,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.7946,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 1.8389166072701353,
2008
+ "grad_norm": 5.59375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.7822,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 1.8460441910192444,
2015
+ "grad_norm": 6.59375,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.7763,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 1.8531717747683536,
2022
+ "grad_norm": 5.90625,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.8152,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 1.8602993585174625,
2029
+ "grad_norm": 5.53125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.8451,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 1.8674269422665715,
2036
+ "grad_norm": 4.6875,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.7819,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 1.8745545260156806,
2043
+ "grad_norm": 5.3125,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.794,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 1.8816821097647898,
2050
+ "grad_norm": 4.8125,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.7435,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 1.888809693513899,
2057
+ "grad_norm": 6.09375,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.8852,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 1.8959372772630079,
2064
+ "grad_norm": 5.46875,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.7538,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 1.9030648610121168,
2071
+ "grad_norm": 5.3125,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.7795,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 1.910192444761226,
2078
+ "grad_norm": 5.65625,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.7887,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 1.9173200285103351,
2085
+ "grad_norm": 6.09375,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.6838,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 1.924447612259444,
2092
+ "grad_norm": 6.4375,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.7495,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 1.931575196008553,
2099
+ "grad_norm": 5.34375,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.7499,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 1.9387027797576621,
2106
+ "grad_norm": 5.21875,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.8243,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 1.9458303635067713,
2113
+ "grad_norm": 5.34375,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.8128,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 1.9529579472558802,
2120
+ "grad_norm": 5.71875,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.7626,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 1.9600855310049892,
2127
+ "grad_norm": 4.96875,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.7724,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 1.9672131147540983,
2134
+ "grad_norm": 5.59375,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.7741,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 1.9743406985032075,
2141
+ "grad_norm": 6.3125,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.7482,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 1.9814682822523164,
2148
+ "grad_norm": 5.6875,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.7981,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 1.9885958660014256,
2155
+ "grad_norm": 4.875,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.7901,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 1.9957234497505345,
2162
+ "grad_norm": 5.96875,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.7576,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.0028510334996437,
2169
+ "grad_norm": 5.4375,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.7789,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.009978617248753,
2176
+ "grad_norm": 4.96875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.8136,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.0171062009978615,
2183
+ "grad_norm": 5.1875,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.813,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.0242337847469707,
2190
+ "grad_norm": 5.96875,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.7948,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.03136136849608,
2197
+ "grad_norm": 5.09375,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6773,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.038488952245189,
2204
+ "grad_norm": 5.03125,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.6645,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.0456165359942977,
2211
+ "grad_norm": 5.34375,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.7572,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.052744119743407,
2218
+ "grad_norm": 4.59375,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.7588,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.059871703492516,
2225
+ "grad_norm": 5.625,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.7294,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.066999287241625,
2232
+ "grad_norm": 5.09375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.7621,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.0741268709907343,
2239
+ "grad_norm": 6.65625,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.8967,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.081254454739843,
2246
+ "grad_norm": 5.3125,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.84,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.088382038488952,
2253
+ "grad_norm": 4.375,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.7584,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.0955096222380614,
2260
+ "grad_norm": 5.21875,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.8475,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.1026372059871705,
2267
+ "grad_norm": 6.40625,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.7067,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.1097647897362792,
2274
+ "grad_norm": 6.1875,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.8023,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.1168923734853884,
2281
+ "grad_norm": 5.03125,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.7074,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.1240199572344975,
2288
+ "grad_norm": 7.5,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.7498,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.1311475409836067,
2295
+ "grad_norm": 7.0625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.7249,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.1382751247327154,
2302
+ "grad_norm": 5.53125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.7907,
2305
+ "step": 3000
2306
+ },
2307
+ {
2308
+ "epoch": 2.1382751247327154,
2309
+ "eval/nano_beir.arguana": 0.4886028250717679,
2310
+ "step": 3000
2311
+ },
2312
+ {
2313
+ "epoch": 2.1382751247327154,
2314
+ "eval/nano_beir.climate_fever": 0.2710241560666136,
2315
+ "step": 3000
2316
+ },
2317
+ {
2318
+ "epoch": 2.1382751247327154,
2319
+ "eval/nano_beir.dbpedia_entity": 0.3285043023237046,
2320
+ "step": 3000
2321
+ },
2322
+ {
2323
+ "epoch": 2.1382751247327154,
2324
+ "eval/nano_beir.fever": 0.6331373355635795,
2325
+ "step": 3000
2326
+ },
2327
+ {
2328
+ "epoch": 2.1382751247327154,
2329
+ "eval/nano_beir.fiqa": 0.4150357832110943,
2330
+ "step": 3000
2331
+ },
2332
+ {
2333
+ "epoch": 2.1382751247327154,
2334
+ "eval/nano_beir.hotpotqa": 0.68158322360907,
2335
+ "step": 3000
2336
+ },
2337
+ {
2338
+ "epoch": 2.1382751247327154,
2339
+ "eval/nano_beir.nfcorpus": 0.3531375464357257,
2340
+ "step": 3000
2341
+ },
2342
+ {
2343
+ "epoch": 2.1382751247327154,
2344
+ "eval/nano_beir.nq": 0.41568885251388027,
2345
+ "step": 3000
2346
+ },
2347
+ {
2348
+ "epoch": 2.1382751247327154,
2349
+ "eval/nano_beir.quora": 0.8349377083144377,
2350
+ "step": 3000
2351
+ },
2352
+ {
2353
+ "epoch": 2.1382751247327154,
2354
+ "eval/nano_beir.scidocs": 0.24315533084930935,
2355
+ "step": 3000
2356
+ },
2357
+ {
2358
+ "epoch": 2.1382751247327154,
2359
+ "eval/nano_beir.scifact": 0.7290861351959168,
2360
+ "step": 3000
2361
+ },
2362
+ {
2363
+ "epoch": 2.1382751247327154,
2364
+ "eval/nano_beir.webis_touche2020": 0.28783156872457877,
2365
+ "step": 3000
2366
+ },
2367
+ {
2368
+ "epoch": 2.1382751247327154,
2369
+ "eval/avg": 0.47347706398997325,
2370
+ "step": 3000
2371
+ },
2372
+ {
2373
+ "epoch": 2.1454027084818246,
2374
+ "grad_norm": 5.75,
2375
+ "learning_rate": 4.100837805544279e-05,
2376
+ "loss": 0.7733,
2377
+ "step": 3010
2378
+ },
2379
+ {
2380
+ "epoch": 2.1525302922309337,
2381
+ "grad_norm": 4.28125,
2382
+ "learning_rate": 4.0665224844564053e-05,
2383
+ "loss": 0.7141,
2384
+ "step": 3020
2385
+ },
2386
+ {
2387
+ "epoch": 2.159657875980043,
2388
+ "grad_norm": 6.8125,
2389
+ "learning_rate": 4.032252659745699e-05,
2390
+ "loss": 0.6352,
2391
+ "step": 3030
2392
+ },
2393
+ {
2394
+ "epoch": 2.1667854597291516,
2395
+ "grad_norm": 4.75,
2396
+ "learning_rate": 3.9980300016750694e-05,
2397
+ "loss": 0.8184,
2398
+ "step": 3040
2399
+ },
2400
+ {
2401
+ "epoch": 2.1739130434782608,
2402
+ "grad_norm": 5.0625,
2403
+ "learning_rate": 3.963856178208588e-05,
2404
+ "loss": 0.7785,
2405
+ "step": 3050
2406
+ },
2407
+ {
2408
+ "epoch": 2.18104062722737,
2409
+ "grad_norm": 4.875,
2410
+ "learning_rate": 3.9297328549302e-05,
2411
+ "loss": 0.7565,
2412
+ "step": 3060
2413
+ },
2414
+ {
2415
+ "epoch": 2.188168210976479,
2416
+ "grad_norm": 5.90625,
2417
+ "learning_rate": 3.895661694962542e-05,
2418
+ "loss": 0.7608,
2419
+ "step": 3070
2420
+ },
2421
+ {
2422
+ "epoch": 2.1952957947255882,
2423
+ "grad_norm": 5.90625,
2424
+ "learning_rate": 3.8616443588858805e-05,
2425
+ "loss": 0.7354,
2426
+ "step": 3080
2427
+ },
2428
+ {
2429
+ "epoch": 2.202423378474697,
2430
+ "grad_norm": 5.28125,
2431
+ "learning_rate": 3.827682504657187e-05,
2432
+ "loss": 0.7828,
2433
+ "step": 3090
2434
+ },
2435
+ {
2436
+ "epoch": 2.209550962223806,
2437
+ "grad_norm": 7.1875,
2438
+ "learning_rate": 3.7937777875293244e-05,
2439
+ "loss": 0.7934,
2440
+ "step": 3100
2441
+ },
2442
+ {
2443
+ "epoch": 2.2166785459729152,
2444
+ "grad_norm": 3.859375,
2445
+ "learning_rate": 3.759931859970374e-05,
2446
+ "loss": 0.6331,
2447
+ "step": 3110
2448
+ },
2449
+ {
2450
+ "epoch": 2.2238061297220244,
2451
+ "grad_norm": 5.09375,
2452
+ "learning_rate": 3.72614637158309e-05,
2453
+ "loss": 0.7351,
2454
+ "step": 3120
2455
+ },
2456
+ {
2457
+ "epoch": 2.230933713471133,
2458
+ "grad_norm": 6.15625,
2459
+ "learning_rate": 3.692422969024516e-05,
2460
+ "loss": 0.7856,
2461
+ "step": 3130
2462
+ },
2463
+ {
2464
+ "epoch": 2.2380612972202423,
2465
+ "grad_norm": 5.5,
2466
+ "learning_rate": 3.6587632959257165e-05,
2467
+ "loss": 0.7819,
2468
+ "step": 3140
2469
+ },
2470
+ {
2471
+ "epoch": 2.2451888809693514,
2472
+ "grad_norm": 5.09375,
2473
+ "learning_rate": 3.625168992811671e-05,
2474
+ "loss": 0.7397,
2475
+ "step": 3150
2476
+ },
2477
+ {
2478
+ "epoch": 2.2523164647184606,
2479
+ "grad_norm": 5.21875,
2480
+ "learning_rate": 3.591641697021317e-05,
2481
+ "loss": 0.7588,
2482
+ "step": 3160
2483
+ },
2484
+ {
2485
+ "epoch": 2.2594440484675697,
2486
+ "grad_norm": 5.5625,
2487
+ "learning_rate": 3.558183042627755e-05,
2488
+ "loss": 0.751,
2489
+ "step": 3170
2490
+ },
2491
+ {
2492
+ "epoch": 2.2665716322166785,
2493
+ "grad_norm": 5.46875,
2494
+ "learning_rate": 3.524794660358593e-05,
2495
+ "loss": 0.8684,
2496
+ "step": 3180
2497
+ },
2498
+ {
2499
+ "epoch": 2.2736992159657876,
2500
+ "grad_norm": 4.5625,
2501
+ "learning_rate": 3.491478177516484e-05,
2502
+ "loss": 0.7863,
2503
+ "step": 3190
2504
+ },
2505
+ {
2506
+ "epoch": 2.2808267997148968,
2507
+ "grad_norm": 5.25,
2508
+ "learning_rate": 3.4582352178997935e-05,
2509
+ "loss": 0.7262,
2510
+ "step": 3200
2511
+ },
2512
+ {
2513
+ "epoch": 2.2879543834640055,
2514
+ "grad_norm": 4.0,
2515
+ "learning_rate": 3.425067401723477e-05,
2516
+ "loss": 0.7284,
2517
+ "step": 3210
2518
+ },
2519
+ {
2520
+ "epoch": 2.2950819672131146,
2521
+ "grad_norm": 5.3125,
2522
+ "learning_rate": 3.3919763455401015e-05,
2523
+ "loss": 0.7184,
2524
+ "step": 3220
2525
+ },
2526
+ {
2527
+ "epoch": 2.302209550962224,
2528
+ "grad_norm": 5.21875,
2529
+ "learning_rate": 3.358963662161062e-05,
2530
+ "loss": 0.8039,
2531
+ "step": 3230
2532
+ },
2533
+ {
2534
+ "epoch": 2.309337134711333,
2535
+ "grad_norm": 4.28125,
2536
+ "learning_rate": 3.326030960577972e-05,
2537
+ "loss": 0.7646,
2538
+ "step": 3240
2539
+ },
2540
+ {
2541
+ "epoch": 2.316464718460442,
2542
+ "grad_norm": 5.96875,
2543
+ "learning_rate": 3.293179845884245e-05,
2544
+ "loss": 0.7086,
2545
+ "step": 3250
2546
+ },
2547
+ {
2548
+ "epoch": 2.323592302209551,
2549
+ "grad_norm": 5.6875,
2550
+ "learning_rate": 3.2604119191968654e-05,
2551
+ "loss": 0.8153,
2552
+ "step": 3260
2553
+ },
2554
+ {
2555
+ "epoch": 2.33071988595866,
2556
+ "grad_norm": 6.125,
2557
+ "learning_rate": 3.2277287775783525e-05,
2558
+ "loss": 0.7863,
2559
+ "step": 3270
2560
+ },
2561
+ {
2562
+ "epoch": 2.337847469707769,
2563
+ "grad_norm": 5.90625,
2564
+ "learning_rate": 3.195132013958918e-05,
2565
+ "loss": 0.7569,
2566
+ "step": 3280
2567
+ },
2568
+ {
2569
+ "epoch": 2.3449750534568783,
2570
+ "grad_norm": 4.34375,
2571
+ "learning_rate": 3.162623217058834e-05,
2572
+ "loss": 0.7193,
2573
+ "step": 3290
2574
+ },
2575
+ {
2576
+ "epoch": 2.352102637205987,
2577
+ "grad_norm": 7.15625,
2578
+ "learning_rate": 3.130203971310999e-05,
2579
+ "loss": 0.847,
2580
+ "step": 3300
2581
+ },
2582
+ {
2583
+ "epoch": 2.359230220955096,
2584
+ "grad_norm": 4.6875,
2585
+ "learning_rate": 3.097875856783713e-05,
2586
+ "loss": 0.7884,
2587
+ "step": 3310
2588
+ },
2589
+ {
2590
+ "epoch": 2.3663578047042053,
2591
+ "grad_norm": 4.46875,
2592
+ "learning_rate": 3.065640449103669e-05,
2593
+ "loss": 0.7849,
2594
+ "step": 3320
2595
+ },
2596
+ {
2597
+ "epoch": 2.3734853884533145,
2598
+ "grad_norm": 4.84375,
2599
+ "learning_rate": 3.033499319379163e-05,
2600
+ "loss": 0.7069,
2601
+ "step": 3330
2602
+ },
2603
+ {
2604
+ "epoch": 2.3806129722024236,
2605
+ "grad_norm": 4.28125,
2606
+ "learning_rate": 3.001454034123512e-05,
2607
+ "loss": 0.6801,
2608
+ "step": 3340
2609
+ },
2610
+ {
2611
+ "epoch": 2.3877405559515323,
2612
+ "grad_norm": 5.34375,
2613
+ "learning_rate": 2.9695061551787108e-05,
2614
+ "loss": 0.7755,
2615
+ "step": 3350
2616
+ },
2617
+ {
2618
+ "epoch": 2.3948681397006415,
2619
+ "grad_norm": 4.21875,
2620
+ "learning_rate": 2.9376572396393048e-05,
2621
+ "loss": 0.8031,
2622
+ "step": 3360
2623
+ },
2624
+ {
2625
+ "epoch": 2.4019957234497507,
2626
+ "grad_norm": 5.34375,
2627
+ "learning_rate": 2.9059088397765088e-05,
2628
+ "loss": 0.8501,
2629
+ "step": 3370
2630
+ },
2631
+ {
2632
+ "epoch": 2.4091233071988594,
2633
+ "grad_norm": 4.4375,
2634
+ "learning_rate": 2.874262502962537e-05,
2635
+ "loss": 0.8451,
2636
+ "step": 3380
2637
+ },
2638
+ {
2639
+ "epoch": 2.4162508909479685,
2640
+ "grad_norm": 4.3125,
2641
+ "learning_rate": 2.8427197715952047e-05,
2642
+ "loss": 0.7452,
2643
+ "step": 3390
2644
+ },
2645
+ {
2646
+ "epoch": 2.4233784746970777,
2647
+ "grad_norm": 4.8125,
2648
+ "learning_rate": 2.811282183022736e-05,
2649
+ "loss": 0.6863,
2650
+ "step": 3400
2651
+ },
2652
+ {
2653
+ "epoch": 2.430506058446187,
2654
+ "grad_norm": 5.5,
2655
+ "learning_rate": 2.7799512694688468e-05,
2656
+ "loss": 0.811,
2657
+ "step": 3410
2658
+ },
2659
+ {
2660
+ "epoch": 2.437633642195296,
2661
+ "grad_norm": 4.375,
2662
+ "learning_rate": 2.7487285579580637e-05,
2663
+ "loss": 0.7234,
2664
+ "step": 3420
2665
+ },
2666
+ {
2667
+ "epoch": 2.4447612259444047,
2668
+ "grad_norm": 5.0625,
2669
+ "learning_rate": 2.7176155702412937e-05,
2670
+ "loss": 0.8367,
2671
+ "step": 3430
2672
+ },
2673
+ {
2674
+ "epoch": 2.451888809693514,
2675
+ "grad_norm": 5.90625,
2676
+ "learning_rate": 2.686613822721666e-05,
2677
+ "loss": 0.6122,
2678
+ "step": 3440
2679
+ },
2680
+ {
2681
+ "epoch": 2.459016393442623,
2682
+ "grad_norm": 4.4375,
2683
+ "learning_rate": 2.6557248263806174e-05,
2684
+ "loss": 0.7393,
2685
+ "step": 3450
2686
+ },
2687
+ {
2688
+ "epoch": 2.466143977191732,
2689
+ "grad_norm": 5.0,
2690
+ "learning_rate": 2.6249500867042522e-05,
2691
+ "loss": 0.7783,
2692
+ "step": 3460
2693
+ },
2694
+ {
2695
+ "epoch": 2.473271560940841,
2696
+ "grad_norm": 5.1875,
2697
+ "learning_rate": 2.5942911036099655e-05,
2698
+ "loss": 0.6654,
2699
+ "step": 3470
2700
+ },
2701
+ {
2702
+ "epoch": 2.48039914468995,
2703
+ "grad_norm": 4.53125,
2704
+ "learning_rate": 2.5637493713733374e-05,
2705
+ "loss": 0.752,
2706
+ "step": 3480
2707
+ },
2708
+ {
2709
+ "epoch": 2.487526728439059,
2710
+ "grad_norm": 5.1875,
2711
+ "learning_rate": 2.533326378555314e-05,
2712
+ "loss": 0.7761,
2713
+ "step": 3490
2714
+ },
2715
+ {
2716
+ "epoch": 2.4946543121881684,
2717
+ "grad_norm": 5.53125,
2718
+ "learning_rate": 2.5030236079296444e-05,
2719
+ "loss": 0.7594,
2720
+ "step": 3500
2721
+ },
2722
+ {
2723
+ "epoch": 2.5017818959372775,
2724
+ "grad_norm": 5.6875,
2725
+ "learning_rate": 2.4728425364106135e-05,
2726
+ "loss": 0.7519,
2727
+ "step": 3510
2728
+ },
2729
+ {
2730
+ "epoch": 2.5089094796863862,
2731
+ "grad_norm": 5.1875,
2732
+ "learning_rate": 2.442784634981071e-05,
2733
+ "loss": 0.7527,
2734
+ "step": 3520
2735
+ },
2736
+ {
2737
+ "epoch": 2.5160370634354954,
2738
+ "grad_norm": 4.28125,
2739
+ "learning_rate": 2.412851368620726e-05,
2740
+ "loss": 0.6595,
2741
+ "step": 3530
2742
+ },
2743
+ {
2744
+ "epoch": 2.5231646471846045,
2745
+ "grad_norm": 4.53125,
2746
+ "learning_rate": 2.383044196234753e-05,
2747
+ "loss": 0.7621,
2748
+ "step": 3540
2749
+ },
2750
+ {
2751
+ "epoch": 2.5302922309337132,
2752
+ "grad_norm": 4.28125,
2753
+ "learning_rate": 2.3533645705826806e-05,
2754
+ "loss": 0.7766,
2755
+ "step": 3550
2756
+ },
2757
+ {
2758
+ "epoch": 2.5374198146828224,
2759
+ "grad_norm": 4.625,
2760
+ "learning_rate": 2.323813938207593e-05,
2761
+ "loss": 0.7717,
2762
+ "step": 3560
2763
+ },
2764
+ {
2765
+ "epoch": 2.5445473984319316,
2766
+ "grad_norm": 5.625,
2767
+ "learning_rate": 2.294393739365621e-05,
2768
+ "loss": 0.6604,
2769
+ "step": 3570
2770
+ },
2771
+ {
2772
+ "epoch": 2.5516749821810407,
2773
+ "grad_norm": 4.03125,
2774
+ "learning_rate": 2.265105407955752e-05,
2775
+ "loss": 0.7614,
2776
+ "step": 3580
2777
+ },
2778
+ {
2779
+ "epoch": 2.55880256593015,
2780
+ "grad_norm": 5.59375,
2781
+ "learning_rate": 2.235950371449938e-05,
2782
+ "loss": 0.7702,
2783
+ "step": 3590
2784
+ },
2785
+ {
2786
+ "epoch": 2.5659301496792586,
2787
+ "grad_norm": 5.96875,
2788
+ "learning_rate": 2.2069300508235275e-05,
2789
+ "loss": 0.8866,
2790
+ "step": 3600
2791
+ },
2792
+ {
2793
+ "epoch": 2.5730577334283677,
2794
+ "grad_norm": 6.21875,
2795
+ "learning_rate": 2.1780458604860056e-05,
2796
+ "loss": 0.8359,
2797
+ "step": 3610
2798
+ },
2799
+ {
2800
+ "epoch": 2.580185317177477,
2801
+ "grad_norm": 4.75,
2802
+ "learning_rate": 2.14929920821206e-05,
2803
+ "loss": 0.7036,
2804
+ "step": 3620
2805
+ },
2806
+ {
2807
+ "epoch": 2.5873129009265856,
2808
+ "grad_norm": 6.25,
2809
+ "learning_rate": 2.1206914950729672e-05,
2810
+ "loss": 0.7946,
2811
+ "step": 3630
2812
+ },
2813
+ {
2814
+ "epoch": 2.5944404846756948,
2815
+ "grad_norm": 8.1875,
2816
+ "learning_rate": 2.0922241153683066e-05,
2817
+ "loss": 0.7035,
2818
+ "step": 3640
2819
+ },
2820
+ {
2821
+ "epoch": 2.601568068424804,
2822
+ "grad_norm": 5.0625,
2823
+ "learning_rate": 2.0638984565580023e-05,
2824
+ "loss": 0.7456,
2825
+ "step": 3650
2826
+ },
2827
+ {
2828
+ "epoch": 2.608695652173913,
2829
+ "grad_norm": 5.6875,
2830
+ "learning_rate": 2.035715899194704e-05,
2831
+ "loss": 0.7462,
2832
+ "step": 3660
2833
+ },
2834
+ {
2835
+ "epoch": 2.6158232359230222,
2836
+ "grad_norm": 4.65625,
2837
+ "learning_rate": 2.007677816856498e-05,
2838
+ "loss": 0.7088,
2839
+ "step": 3670
2840
+ },
2841
+ {
2842
+ "epoch": 2.6229508196721314,
2843
+ "grad_norm": 4.84375,
2844
+ "learning_rate": 1.9797855760799612e-05,
2845
+ "loss": 0.721,
2846
+ "step": 3680
2847
+ },
2848
+ {
2849
+ "epoch": 2.63007840342124,
2850
+ "grad_norm": 5.84375,
2851
+ "learning_rate": 1.9520405362935594e-05,
2852
+ "loss": 0.7292,
2853
+ "step": 3690
2854
+ },
2855
+ {
2856
+ "epoch": 2.6372059871703493,
2857
+ "grad_norm": 5.0,
2858
+ "learning_rate": 1.9244440497513893e-05,
2859
+ "loss": 0.7031,
2860
+ "step": 3700
2861
+ },
2862
+ {
2863
+ "epoch": 2.6443335709194584,
2864
+ "grad_norm": 4.3125,
2865
+ "learning_rate": 1.896997461467272e-05,
2866
+ "loss": 0.696,
2867
+ "step": 3710
2868
+ },
2869
+ {
2870
+ "epoch": 2.651461154668567,
2871
+ "grad_norm": 4.875,
2872
+ "learning_rate": 1.869702109149199e-05,
2873
+ "loss": 0.749,
2874
+ "step": 3720
2875
+ },
2876
+ {
2877
+ "epoch": 2.6585887384176763,
2878
+ "grad_norm": 6.21875,
2879
+ "learning_rate": 1.842559323134136e-05,
2880
+ "loss": 0.7282,
2881
+ "step": 3730
2882
+ },
2883
+ {
2884
+ "epoch": 2.6657163221667854,
2885
+ "grad_norm": 5.0625,
2886
+ "learning_rate": 1.8155704263231776e-05,
2887
+ "loss": 0.6514,
2888
+ "step": 3740
2889
+ },
2890
+ {
2891
+ "epoch": 2.6728439059158946,
2892
+ "grad_norm": 4.9375,
2893
+ "learning_rate": 1.788736734117078e-05,
2894
+ "loss": 0.7496,
2895
+ "step": 3750
2896
+ },
2897
+ {
2898
+ "epoch": 2.6799714896650038,
2899
+ "grad_norm": 4.5,
2900
+ "learning_rate": 1.762059554352143e-05,
2901
+ "loss": 0.8261,
2902
+ "step": 3760
2903
+ },
2904
+ {
2905
+ "epoch": 2.6870990734141125,
2906
+ "grad_norm": 4.8125,
2907
+ "learning_rate": 1.7355401872364758e-05,
2908
+ "loss": 0.8161,
2909
+ "step": 3770
2910
+ },
2911
+ {
2912
+ "epoch": 2.6942266571632216,
2913
+ "grad_norm": 5.125,
2914
+ "learning_rate": 1.709179925286617e-05,
2915
+ "loss": 0.682,
2916
+ "step": 3780
2917
+ },
2918
+ {
2919
+ "epoch": 2.701354240912331,
2920
+ "grad_norm": 4.90625,
2921
+ "learning_rate": 1.6829800532645447e-05,
2922
+ "loss": 0.7542,
2923
+ "step": 3790
2924
+ },
2925
+ {
2926
+ "epoch": 2.7084818246614395,
2927
+ "grad_norm": 6.46875,
2928
+ "learning_rate": 1.6569418481150595e-05,
2929
+ "loss": 0.8126,
2930
+ "step": 3800
2931
+ },
2932
+ {
2933
+ "epoch": 2.7156094084105487,
2934
+ "grad_norm": 4.6875,
2935
+ "learning_rate": 1.6310665789035467e-05,
2936
+ "loss": 0.7445,
2937
+ "step": 3810
2938
+ },
2939
+ {
2940
+ "epoch": 2.722736992159658,
2941
+ "grad_norm": 4.9375,
2942
+ "learning_rate": 1.605355506754121e-05,
2943
+ "loss": 0.8296,
2944
+ "step": 3820
2945
+ },
2946
+ {
2947
+ "epoch": 2.729864575908767,
2948
+ "grad_norm": 4.84375,
2949
+ "learning_rate": 1.5798098847881666e-05,
2950
+ "loss": 0.6684,
2951
+ "step": 3830
2952
+ },
2953
+ {
2954
+ "epoch": 2.736992159657876,
2955
+ "grad_norm": 4.5625,
2956
+ "learning_rate": 1.554430958063259e-05,
2957
+ "loss": 0.7063,
2958
+ "step": 3840
2959
+ },
2960
+ {
2961
+ "epoch": 2.7441197434069853,
2962
+ "grad_norm": 4.625,
2963
+ "learning_rate": 1.529219963512481e-05,
2964
+ "loss": 0.6668,
2965
+ "step": 3850
2966
+ },
2967
+ {
2968
+ "epoch": 2.751247327156094,
2969
+ "grad_norm": 5.21875,
2970
+ "learning_rate": 1.5041781298841423e-05,
2971
+ "loss": 0.7464,
2972
+ "step": 3860
2973
+ },
2974
+ {
2975
+ "epoch": 2.758374910905203,
2976
+ "grad_norm": 4.78125,
2977
+ "learning_rate": 1.4793066776818842e-05,
2978
+ "loss": 0.7625,
2979
+ "step": 3870
2980
+ },
2981
+ {
2982
+ "epoch": 2.7655024946543123,
2983
+ "grad_norm": 5.625,
2984
+ "learning_rate": 1.4546068191051987e-05,
2985
+ "loss": 0.738,
2986
+ "step": 3880
2987
+ },
2988
+ {
2989
+ "epoch": 2.772630078403421,
2990
+ "grad_norm": 6.125,
2991
+ "learning_rate": 1.4300797579903474e-05,
2992
+ "loss": 0.8185,
2993
+ "step": 3890
2994
+ },
2995
+ {
2996
+ "epoch": 2.77975766215253,
2997
+ "grad_norm": 5.4375,
2998
+ "learning_rate": 1.4057266897516841e-05,
2999
+ "loss": 0.7021,
3000
+ "step": 3900
3001
+ },
3002
+ {
3003
+ "epoch": 2.7868852459016393,
3004
+ "grad_norm": 6.78125,
3005
+ "learning_rate": 1.3815488013233984e-05,
3006
+ "loss": 0.8398,
3007
+ "step": 3910
3008
+ },
3009
+ {
3010
+ "epoch": 2.7940128296507485,
3011
+ "grad_norm": 6.03125,
3012
+ "learning_rate": 1.3575472711016635e-05,
3013
+ "loss": 0.8821,
3014
+ "step": 3920
3015
+ },
3016
+ {
3017
+ "epoch": 2.8011404133998576,
3018
+ "grad_norm": 4.75,
3019
+ "learning_rate": 1.3337232688872009e-05,
3020
+ "loss": 0.7238,
3021
+ "step": 3930
3022
+ },
3023
+ {
3024
+ "epoch": 2.8082679971489664,
3025
+ "grad_norm": 5.1875,
3026
+ "learning_rate": 1.3100779558282672e-05,
3027
+ "loss": 0.6887,
3028
+ "step": 3940
3029
+ },
3030
+ {
3031
+ "epoch": 2.8153955808980755,
3032
+ "grad_norm": 6.03125,
3033
+ "learning_rate": 1.2866124843640615e-05,
3034
+ "loss": 0.7804,
3035
+ "step": 3950
3036
+ },
3037
+ {
3038
+ "epoch": 2.8225231646471847,
3039
+ "grad_norm": 5.78125,
3040
+ "learning_rate": 1.2633279981685608e-05,
3041
+ "loss": 0.871,
3042
+ "step": 3960
3043
+ },
3044
+ {
3045
+ "epoch": 2.8296507483962934,
3046
+ "grad_norm": 4.40625,
3047
+ "learning_rate": 1.240225632094773e-05,
3048
+ "loss": 0.764,
3049
+ "step": 3970
3050
+ },
3051
+ {
3052
+ "epoch": 2.8367783321454025,
3053
+ "grad_norm": 4.65625,
3054
+ "learning_rate": 1.217306512119425e-05,
3055
+ "loss": 0.7754,
3056
+ "step": 3980
3057
+ },
3058
+ {
3059
+ "epoch": 2.8439059158945117,
3060
+ "grad_norm": 5.71875,
3061
+ "learning_rate": 1.194571755288092e-05,
3062
+ "loss": 0.7185,
3063
+ "step": 3990
3064
+ },
3065
+ {
3066
+ "epoch": 2.851033499643621,
3067
+ "grad_norm": 4.9375,
3068
+ "learning_rate": 1.1720224696607474e-05,
3069
+ "loss": 0.6823,
3070
+ "step": 4000
3071
+ }
3072
+ ],
3073
+ "logging_steps": 10,
3074
+ "max_steps": 5000,
3075
+ "num_input_tokens_seen": 0,
3076
+ "num_train_epochs": 4,
3077
+ "save_steps": 1000,
3078
+ "stateful_callbacks": {
3079
+ "TrainerControl": {
3080
+ "args": {
3081
+ "should_epoch_stop": false,
3082
+ "should_evaluate": false,
3083
+ "should_log": false,
3084
+ "should_save": true,
3085
+ "should_training_stop": false
3086
+ },
3087
+ "attributes": {}
3088
+ }
3089
+ },
3090
+ "total_flos": 0.0,
3091
+ "train_batch_size": 16,
3092
+ "trial_name": null,
3093
+ "trial_params": null
3094
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-4000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1afec67e7d18e7478623434fcf8c1bcc7719c01c4f7874089687d7138953afff
3
+ size 6097
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e1a5059a50ad8539351fafc2820f9d48123be1a23a2be6777263c5223144834
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46afb89b2e44d638dea07819f23ad7f4dcae4b7942aab09479a61a4fa5373638
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c12cceaffe8acec04189ba71dbb81f100dd5cd03b86390253117b3fab2911d
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24e46cf0599262f5ee2ba9ec6bddad7d3e8feb965b424cc9c4dc4563771b1e2
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5af6ed530327c4af91aa04cc19a7523ed8295478ca34bb5f3124ff88cd33f77
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74c82a4770595a28e603ecf5353ab454c0fb9b0e184187b20ebdc4685217f3a5
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23766ee8d5dbc082836e1a591a2e0b269bc9bd5027b36062367e8f61a621f007
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.filtered.b64_n512.1e-4/checkpoint-5000/trainer_state.json ADDED
@@ -0,0 +1,3859 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.563791874554526,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 5000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007127583749109052,
79
+ "grad_norm": 11.9375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.5314,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014255167498218105,
86
+ "grad_norm": 11.125,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.4333,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021382751247327157,
93
+ "grad_norm": 22.75,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.6002,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.02851033499643621,
100
+ "grad_norm": 13.3125,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.5692,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03563791874554526,
107
+ "grad_norm": 14.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.2221,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042765502494654314,
114
+ "grad_norm": 12.4375,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.3981,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04989308624376337,
121
+ "grad_norm": 11.375,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.3844,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05702066999287242,
128
+ "grad_norm": 12.375,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.6756,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06414825374198146,
135
+ "grad_norm": 9.8125,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3874,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07127583749109052,
142
+ "grad_norm": 11.5,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.3705,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07840342124019957,
149
+ "grad_norm": 10.0,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3897,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08553100498930863,
156
+ "grad_norm": 11.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.4046,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09265858873841767,
163
+ "grad_norm": 8.25,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1084,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09978617248752673,
170
+ "grad_norm": 7.96875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.1872,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10691375623663578,
177
+ "grad_norm": 10.8125,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.4332,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11404133998574484,
184
+ "grad_norm": 12.8125,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2686,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12116892373485388,
191
+ "grad_norm": 12.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 0.9751,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12829650748396293,
198
+ "grad_norm": 9.4375,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1097,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.13542409123307197,
205
+ "grad_norm": 10.6875,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 1.0446,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14255167498218105,
212
+ "grad_norm": 8.625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.2052,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.1496792587312901,
219
+ "grad_norm": 9.5625,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 1.0325,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15680684248039914,
226
+ "grad_norm": 9.1875,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9825,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16393442622950818,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0758,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17106200997861726,
240
+ "grad_norm": 7.9375,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.076,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1781895937277263,
247
+ "grad_norm": 8.25,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0498,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18531717747683535,
254
+ "grad_norm": 7.9375,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9616,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.1924447612259444,
261
+ "grad_norm": 7.71875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 1.0782,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19957234497505347,
268
+ "grad_norm": 8.6875,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.9648,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.2066999287241625,
275
+ "grad_norm": 9.375,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9675,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21382751247327156,
282
+ "grad_norm": 7.53125,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.9063,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2209550962223806,
289
+ "grad_norm": 6.96875,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.8631,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.22808267997148968,
296
+ "grad_norm": 7.28125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.9451,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23521026372059872,
303
+ "grad_norm": 8.9375,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 1.0807,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24233784746970777,
310
+ "grad_norm": 7.09375,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9637,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.2494654312188168,
317
+ "grad_norm": 6.25,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9216,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25659301496792586,
324
+ "grad_norm": 6.34375,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.9588,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.2637205987170349,
331
+ "grad_norm": 6.65625,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8696,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.27084818246614395,
338
+ "grad_norm": 6.0625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.951,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.27797576621525305,
345
+ "grad_norm": 6.6875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.9381,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.2851033499643621,
352
+ "grad_norm": 7.3125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.9314,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29223093371347114,
359
+ "grad_norm": 5.84375,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9212,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2993585174625802,
366
+ "grad_norm": 6.90625,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.8897,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.30648610121168923,
373
+ "grad_norm": 5.46875,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.9813,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.3136136849607983,
380
+ "grad_norm": 4.46875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.9023,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3207412687099073,
387
+ "grad_norm": 5.8125,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.882,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.32786885245901637,
394
+ "grad_norm": 6.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9943,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33499643620812547,
401
+ "grad_norm": 12.0,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8606,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.3421240199572345,
408
+ "grad_norm": 6.125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.9137,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34925160370634356,
415
+ "grad_norm": 6.25,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8677,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3563791874554526,
422
+ "grad_norm": 5.46875,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.9161,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.36350677120456165,
429
+ "grad_norm": 6.125,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.9176,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.3706343549536707,
436
+ "grad_norm": 6.34375,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.871,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.37776193870277974,
443
+ "grad_norm": 6.28125,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.9097,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.3848895224518888,
450
+ "grad_norm": 5.1875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8721,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.3920171062009979,
457
+ "grad_norm": 7.15625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.9282,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39914468995010693,
464
+ "grad_norm": 5.375,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8213,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.406272273699216,
471
+ "grad_norm": 7.28125,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8527,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.413399857448325,
478
+ "grad_norm": 5.75,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.944,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42052744119743407,
485
+ "grad_norm": 6.5625,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.9269,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.4276550249465431,
492
+ "grad_norm": 5.53125,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.8261,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43478260869565216,
499
+ "grad_norm": 6.4375,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 1.0049,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.4419101924447612,
506
+ "grad_norm": 6.65625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.9438,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44903777619387025,
513
+ "grad_norm": 5.9375,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.9527,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.45616535994297935,
520
+ "grad_norm": 6.375,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.8917,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.4632929436920884,
527
+ "grad_norm": 4.625,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8554,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47042052744119744,
534
+ "grad_norm": 5.59375,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.978,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.4775481111903065,
541
+ "grad_norm": 5.0625,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.774,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48467569493941554,
548
+ "grad_norm": 5.1875,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8388,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4918032786885246,
555
+ "grad_norm": 4.8125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.8888,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4989308624376336,
562
+ "grad_norm": 6.03125,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.7606,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5060584461867427,
569
+ "grad_norm": 5.1875,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.9804,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5131860299358517,
576
+ "grad_norm": 6.5625,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.8832,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5203136136849608,
583
+ "grad_norm": 5.0,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.8385,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5274411974340698,
590
+ "grad_norm": 6.78125,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.9106,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5345687811831789,
597
+ "grad_norm": 6.3125,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.807,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5416963649322879,
604
+ "grad_norm": 4.0,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.7932,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5488239486813971,
611
+ "grad_norm": 5.53125,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.7533,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5559515324305061,
618
+ "grad_norm": 5.125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.8676,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5630791161796151,
625
+ "grad_norm": 5.0625,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.8783,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5702066999287242,
632
+ "grad_norm": 5.5625,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.9352,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5773342836778332,
639
+ "grad_norm": 4.8125,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8814,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5844618674269423,
646
+ "grad_norm": 5.96875,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.8252,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5915894511760513,
653
+ "grad_norm": 4.96875,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8695,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5987170349251604,
660
+ "grad_norm": 5.21875,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8821,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6058446186742694,
667
+ "grad_norm": 6.375,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.8968,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6129722024233785,
674
+ "grad_norm": 4.59375,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.9109,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6200997861724875,
681
+ "grad_norm": 6.625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.8701,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6272273699215966,
688
+ "grad_norm": 5.71875,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.8649,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6343549536707056,
695
+ "grad_norm": 5.25,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8578,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6414825374198146,
702
+ "grad_norm": 4.9375,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.9067,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6486101211689237,
709
+ "grad_norm": 6.8125,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.7862,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6557377049180327,
716
+ "grad_norm": 5.875,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.8275,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6628652886671419,
723
+ "grad_norm": 5.6875,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.9155,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6699928724162509,
730
+ "grad_norm": 5.6875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.8526,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.67712045616536,
737
+ "grad_norm": 4.9375,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8123,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.684248039914469,
744
+ "grad_norm": 5.4375,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.7986,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6913756236635781,
751
+ "grad_norm": 5.6875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.9203,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6985032074126871,
758
+ "grad_norm": 5.78125,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.7805,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7056307911617962,
765
+ "grad_norm": 5.4375,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8683,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7127583749109052,
772
+ "grad_norm": 5.46875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8515,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7127583749109052,
779
+ "eval/nano_beir.arguana": 0.49904762134680597,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7127583749109052,
784
+ "eval/nano_beir.climate_fever": 0.2710368630075979,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7127583749109052,
789
+ "eval/nano_beir.dbpedia_entity": 0.32627448935866016,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7127583749109052,
794
+ "eval/nano_beir.fever": 0.6251056136949217,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7127583749109052,
799
+ "eval/nano_beir.fiqa": 0.4264627976119544,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7127583749109052,
804
+ "eval/nano_beir.hotpotqa": 0.67890984670666,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7127583749109052,
809
+ "eval/nano_beir.nfcorpus": 0.33575253271342886,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7127583749109052,
814
+ "eval/nano_beir.nq": 0.41529049449791683,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7127583749109052,
819
+ "eval/nano_beir.quora": 0.8400060722791932,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7127583749109052,
824
+ "eval/nano_beir.scidocs": 0.25236583848912914,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7127583749109052,
829
+ "eval/nano_beir.scifact": 0.7248783076751154,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7127583749109052,
834
+ "eval/nano_beir.webis_touche2020": 0.29804501841705877,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7127583749109052,
839
+ "eval/avg": 0.4744312913165369,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7198859586600143,
844
+ "grad_norm": 5.71875,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8998,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7270135424091233,
851
+ "grad_norm": 5.34375,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8763,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7341411261582323,
858
+ "grad_norm": 4.875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8851,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7412687099073414,
865
+ "grad_norm": 4.40625,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.8247,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7483962936564504,
872
+ "grad_norm": 6.5625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.742,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7555238774055595,
879
+ "grad_norm": 4.6875,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8707,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7626514611546685,
886
+ "grad_norm": 4.8125,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.7698,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7697790449037776,
893
+ "grad_norm": 6.1875,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.9318,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7769066286528866,
900
+ "grad_norm": 4.40625,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.752,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7840342124019958,
907
+ "grad_norm": 6.0,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.9631,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7911617961511048,
914
+ "grad_norm": 4.71875,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.8623,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7982893799002139,
921
+ "grad_norm": 5.84375,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8677,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8054169636493229,
928
+ "grad_norm": 4.03125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.7795,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.812544547398432,
935
+ "grad_norm": 4.125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.8011,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.819672131147541,
942
+ "grad_norm": 4.875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7984,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.82679971489665,
949
+ "grad_norm": 4.84375,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.8049,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8339272986457591,
956
+ "grad_norm": 5.15625,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7839,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8410548823948681,
963
+ "grad_norm": 4.78125,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7809,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8481824661439772,
970
+ "grad_norm": 4.71875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.7688,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8553100498930862,
977
+ "grad_norm": 7.75,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.8638,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8624376336421953,
984
+ "grad_norm": 5.9375,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8409,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8695652173913043,
991
+ "grad_norm": 5.875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.916,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8766928011404134,
998
+ "grad_norm": 5.40625,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.9114,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.8838203848895224,
1005
+ "grad_norm": 4.5,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.8744,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.8909479686386315,
1012
+ "grad_norm": 4.9375,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.8471,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8980755523877405,
1019
+ "grad_norm": 5.21875,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8014,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.9052031361368497,
1026
+ "grad_norm": 5.71875,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.8587,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.9123307198859587,
1033
+ "grad_norm": 5.9375,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8003,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.9194583036350678,
1040
+ "grad_norm": 6.09375,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.7404,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.9265858873841768,
1047
+ "grad_norm": 5.5,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.722,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9337134711332858,
1054
+ "grad_norm": 5.21875,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7005,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9408410548823949,
1061
+ "grad_norm": 6.78125,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.7987,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.9479686386315039,
1068
+ "grad_norm": 5.78125,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7872,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.955096222380613,
1075
+ "grad_norm": 4.875,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.8632,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.962223806129722,
1082
+ "grad_norm": 6.96875,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.8165,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9693513898788311,
1089
+ "grad_norm": 5.0625,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.7802,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9764789736279401,
1096
+ "grad_norm": 5.34375,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.8668,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9836065573770492,
1103
+ "grad_norm": 5.84375,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.7785,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9907341411261582,
1110
+ "grad_norm": 6.25,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8952,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9978617248752673,
1117
+ "grad_norm": 6.34375,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.826,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.0049893086243764,
1124
+ "grad_norm": 6.625,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.9168,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0121168923734853,
1131
+ "grad_norm": 5.625,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8506,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0192444761225945,
1138
+ "grad_norm": 4.875,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.9257,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0263720598717034,
1145
+ "grad_norm": 5.96875,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.8302,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0334996436208126,
1152
+ "grad_norm": 5.375,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.7898,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0406272273699215,
1159
+ "grad_norm": 5.75,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.7682,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0477548111190307,
1166
+ "grad_norm": 4.9375,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.8378,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0548823948681396,
1173
+ "grad_norm": 4.15625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.7113,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0620099786172488,
1180
+ "grad_norm": 5.71875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.7187,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0691375623663577,
1187
+ "grad_norm": 5.21875,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.7843,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0762651461154669,
1194
+ "grad_norm": 5.28125,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.7253,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0833927298645758,
1201
+ "grad_norm": 4.78125,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.8077,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.090520313613685,
1208
+ "grad_norm": 5.8125,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.8465,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0976478973627941,
1215
+ "grad_norm": 4.875,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.7467,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.104775481111903,
1222
+ "grad_norm": 5.0,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.7919,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1119030648610122,
1229
+ "grad_norm": 5.78125,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7399,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1190306486101211,
1236
+ "grad_norm": 5.28125,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.8305,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1261582323592303,
1243
+ "grad_norm": 5.09375,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.8986,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1332858161083392,
1250
+ "grad_norm": 5.65625,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8041,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1404133998574484,
1257
+ "grad_norm": 5.5625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8434,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1475409836065573,
1264
+ "grad_norm": 5.9375,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7681,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1546685673556665,
1271
+ "grad_norm": 4.71875,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.6835,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1617961511047754,
1278
+ "grad_norm": 4.8125,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.8917,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1689237348538846,
1285
+ "grad_norm": 5.25,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8296,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1760513186029935,
1292
+ "grad_norm": 4.90625,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8147,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1831789023521027,
1299
+ "grad_norm": 6.90625,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.7656,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1903064861012118,
1306
+ "grad_norm": 5.28125,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8931,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1974340698503207,
1313
+ "grad_norm": 4.21875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.7844,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2045616535994297,
1320
+ "grad_norm": 5.53125,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7334,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2116892373485388,
1327
+ "grad_norm": 5.28125,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7509,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.218816821097648,
1334
+ "grad_norm": 5.5,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8287,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.225944404846757,
1341
+ "grad_norm": 5.4375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.8383,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.233071988595866,
1348
+ "grad_norm": 6.6875,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7443,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.240199572344975,
1355
+ "grad_norm": 6.0625,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7868,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2473271560940842,
1362
+ "grad_norm": 5.09375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.7662,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2544547398431931,
1369
+ "grad_norm": 4.90625,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7558,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2615823235923023,
1376
+ "grad_norm": 5.34375,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.6894,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2687099073414112,
1383
+ "grad_norm": 5.15625,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7782,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2758374910905204,
1390
+ "grad_norm": 5.34375,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.756,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2829650748396293,
1397
+ "grad_norm": 4.125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.763,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.2900926585887385,
1404
+ "grad_norm": 10.5,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7255,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.2972202423378474,
1411
+ "grad_norm": 4.65625,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7615,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3043478260869565,
1418
+ "grad_norm": 5.15625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8092,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3114754098360657,
1425
+ "grad_norm": 4.71875,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7503,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3186029935851746,
1432
+ "grad_norm": 4.5,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.8121,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.3257305773342836,
1439
+ "grad_norm": 4.78125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.8773,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.3328581610833927,
1446
+ "grad_norm": 5.6875,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.8202,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.3399857448325019,
1453
+ "grad_norm": 6.15625,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.8359,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3471133285816108,
1460
+ "grad_norm": 5.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.9009,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.3542409123307197,
1467
+ "grad_norm": 4.96875,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.7515,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.361368496079829,
1474
+ "grad_norm": 4.96875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.8321,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.368496079828938,
1481
+ "grad_norm": 6.875,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.8211,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.375623663578047,
1488
+ "grad_norm": 3.921875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7751,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.3827512473271562,
1495
+ "grad_norm": 5.53125,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8174,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.389878831076265,
1502
+ "grad_norm": 5.28125,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7058,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3970064148253742,
1509
+ "grad_norm": 6.34375,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7263,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.4041339985744832,
1516
+ "grad_norm": 5.5625,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.7503,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.4112615823235923,
1523
+ "grad_norm": 5.34375,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.87,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4183891660727013,
1530
+ "grad_norm": 4.84375,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7866,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4255167498218104,
1537
+ "grad_norm": 5.0625,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.787,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.4255167498218104,
1544
+ "eval/nano_beir.arguana": 0.45921477926341064,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.4255167498218104,
1549
+ "eval/nano_beir.climate_fever": 0.26026043643827973,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.4255167498218104,
1554
+ "eval/nano_beir.dbpedia_entity": 0.32492910818341564,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.4255167498218104,
1559
+ "eval/nano_beir.fever": 0.6243268043774819,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.4255167498218104,
1564
+ "eval/nano_beir.fiqa": 0.406434177003067,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.4255167498218104,
1569
+ "eval/nano_beir.hotpotqa": 0.6761769995795371,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.4255167498218104,
1574
+ "eval/nano_beir.nfcorpus": 0.3490707285646159,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.4255167498218104,
1579
+ "eval/nano_beir.nq": 0.40743688778055903,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.4255167498218104,
1584
+ "eval/nano_beir.quora": 0.8329214304594091,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.4255167498218104,
1589
+ "eval/nano_beir.scidocs": 0.2444219693805352,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.4255167498218104,
1594
+ "eval/nano_beir.scifact": 0.7389917242185631,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.4255167498218104,
1599
+ "eval/nano_beir.webis_touche2020": 0.2912623613111957,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.4255167498218104,
1604
+ "eval/avg": 0.4679539505466725,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.4326443335709196,
1609
+ "grad_norm": 5.53125,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.873,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.4397719173200285,
1616
+ "grad_norm": 6.34375,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.8482,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.4468995010691375,
1623
+ "grad_norm": 4.875,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.8179,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.4540270848182466,
1630
+ "grad_norm": 4.40625,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.7916,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.4611546685673558,
1637
+ "grad_norm": 4.84375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.7528,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.4682822523164647,
1644
+ "grad_norm": 5.75,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.7749,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.4754098360655736,
1651
+ "grad_norm": 5.625,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.8334,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.4825374198146828,
1658
+ "grad_norm": 7.03125,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.7086,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.489665003563792,
1665
+ "grad_norm": 5.25,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.7891,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.4967925873129009,
1672
+ "grad_norm": 5.0625,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.7186,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.5039201710620098,
1679
+ "grad_norm": 5.28125,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.778,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.511047754811119,
1686
+ "grad_norm": 4.875,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.7799,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.5181753385602281,
1693
+ "grad_norm": 6.09375,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.7962,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.5253029223093373,
1700
+ "grad_norm": 5.84375,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.8263,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.5324305060584462,
1707
+ "grad_norm": 4.78125,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.7541,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.5395580898075552,
1714
+ "grad_norm": 5.1875,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.7617,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.5466856735566643,
1721
+ "grad_norm": 6.28125,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.8576,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.5538132573057735,
1728
+ "grad_norm": 5.6875,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.7801,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.5609408410548824,
1735
+ "grad_norm": 5.96875,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.7256,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.5680684248039913,
1742
+ "grad_norm": 4.03125,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.7574,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.5751960085531005,
1749
+ "grad_norm": 6.65625,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.7539,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.5823235923022096,
1756
+ "grad_norm": 6.25,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.7865,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.5894511760513186,
1763
+ "grad_norm": 4.90625,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.781,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.5965787598004275,
1770
+ "grad_norm": 6.5625,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.9054,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.6037063435495367,
1777
+ "grad_norm": 4.0625,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.7958,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.6108339272986458,
1784
+ "grad_norm": 5.8125,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.7319,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.6179615110477548,
1791
+ "grad_norm": 5.5625,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.7538,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.6250890947968637,
1798
+ "grad_norm": 4.625,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.7781,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.6322166785459729,
1805
+ "grad_norm": 6.21875,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.7636,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.639344262295082,
1812
+ "grad_norm": 5.65625,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.7041,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.6464718460441912,
1819
+ "grad_norm": 4.53125,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.8014,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.6535994297933,
1826
+ "grad_norm": 4.8125,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.7043,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.660727013542409,
1833
+ "grad_norm": 4.78125,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.7739,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.6678545972915182,
1840
+ "grad_norm": 5.0625,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.7807,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.6749821810406273,
1847
+ "grad_norm": 4.84375,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.8122,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.6821097647897363,
1854
+ "grad_norm": 6.25,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.8824,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.6892373485388452,
1861
+ "grad_norm": 4.46875,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.6925,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.6963649322879544,
1868
+ "grad_norm": 5.59375,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.7558,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.7034925160370635,
1875
+ "grad_norm": 5.09375,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.7045,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.7106200997861725,
1882
+ "grad_norm": 5.0625,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.7654,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.7177476835352814,
1889
+ "grad_norm": 4.90625,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.7696,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.7248752672843906,
1896
+ "grad_norm": 5.65625,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.8205,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.7320028510334997,
1903
+ "grad_norm": 6.1875,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.7752,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.7391304347826086,
1910
+ "grad_norm": 4.4375,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.79,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.7462580185317176,
1917
+ "grad_norm": 4.75,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.9032,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.7533856022808267,
1924
+ "grad_norm": 5.09375,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.7631,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.760513186029936,
1931
+ "grad_norm": 4.9375,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.7612,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.767640769779045,
1938
+ "grad_norm": 5.125,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.7827,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.774768353528154,
1945
+ "grad_norm": 4.8125,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.7827,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.781895937277263,
1952
+ "grad_norm": 4.6875,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.7646,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.789023521026372,
1959
+ "grad_norm": 7.3125,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.8215,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.7961511047754812,
1966
+ "grad_norm": 4.84375,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.7416,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.8032786885245902,
1973
+ "grad_norm": 4.3125,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.7577,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.810406272273699,
1980
+ "grad_norm": 4.34375,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.6839,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 1.8175338560228083,
1987
+ "grad_norm": 4.46875,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.719,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 1.8246614397719174,
1994
+ "grad_norm": 5.875,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.7487,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 1.8317890235210263,
2001
+ "grad_norm": 5.40625,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.7946,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 1.8389166072701353,
2008
+ "grad_norm": 5.59375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.7822,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 1.8460441910192444,
2015
+ "grad_norm": 6.59375,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.7763,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 1.8531717747683536,
2022
+ "grad_norm": 5.90625,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.8152,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 1.8602993585174625,
2029
+ "grad_norm": 5.53125,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.8451,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 1.8674269422665715,
2036
+ "grad_norm": 4.6875,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.7819,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 1.8745545260156806,
2043
+ "grad_norm": 5.3125,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.794,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 1.8816821097647898,
2050
+ "grad_norm": 4.8125,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.7435,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 1.888809693513899,
2057
+ "grad_norm": 6.09375,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.8852,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 1.8959372772630079,
2064
+ "grad_norm": 5.46875,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.7538,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 1.9030648610121168,
2071
+ "grad_norm": 5.3125,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.7795,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 1.910192444761226,
2078
+ "grad_norm": 5.65625,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.7887,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 1.9173200285103351,
2085
+ "grad_norm": 6.09375,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.6838,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 1.924447612259444,
2092
+ "grad_norm": 6.4375,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.7495,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 1.931575196008553,
2099
+ "grad_norm": 5.34375,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.7499,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 1.9387027797576621,
2106
+ "grad_norm": 5.21875,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.8243,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 1.9458303635067713,
2113
+ "grad_norm": 5.34375,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.8128,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 1.9529579472558802,
2120
+ "grad_norm": 5.71875,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.7626,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 1.9600855310049892,
2127
+ "grad_norm": 4.96875,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.7724,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 1.9672131147540983,
2134
+ "grad_norm": 5.59375,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.7741,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 1.9743406985032075,
2141
+ "grad_norm": 6.3125,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.7482,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 1.9814682822523164,
2148
+ "grad_norm": 5.6875,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.7981,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 1.9885958660014256,
2155
+ "grad_norm": 4.875,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.7901,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 1.9957234497505345,
2162
+ "grad_norm": 5.96875,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.7576,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.0028510334996437,
2169
+ "grad_norm": 5.4375,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.7789,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.009978617248753,
2176
+ "grad_norm": 4.96875,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.8136,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.0171062009978615,
2183
+ "grad_norm": 5.1875,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.813,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.0242337847469707,
2190
+ "grad_norm": 5.96875,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.7948,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.03136136849608,
2197
+ "grad_norm": 5.09375,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6773,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.038488952245189,
2204
+ "grad_norm": 5.03125,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.6645,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.0456165359942977,
2211
+ "grad_norm": 5.34375,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.7572,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.052744119743407,
2218
+ "grad_norm": 4.59375,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.7588,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.059871703492516,
2225
+ "grad_norm": 5.625,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.7294,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.066999287241625,
2232
+ "grad_norm": 5.09375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.7621,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.0741268709907343,
2239
+ "grad_norm": 6.65625,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.8967,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.081254454739843,
2246
+ "grad_norm": 5.3125,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.84,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.088382038488952,
2253
+ "grad_norm": 4.375,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.7584,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.0955096222380614,
2260
+ "grad_norm": 5.21875,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.8475,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.1026372059871705,
2267
+ "grad_norm": 6.40625,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.7067,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.1097647897362792,
2274
+ "grad_norm": 6.1875,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.8023,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.1168923734853884,
2281
+ "grad_norm": 5.03125,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.7074,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.1240199572344975,
2288
+ "grad_norm": 7.5,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.7498,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.1311475409836067,
2295
+ "grad_norm": 7.0625,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.7249,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.1382751247327154,
2302
+ "grad_norm": 5.53125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.7907,
2305
+ "step": 3000
2306
+ },
2307
+ {
2308
+ "epoch": 2.1382751247327154,
2309
+ "eval/nano_beir.arguana": 0.4886028250717679,
2310
+ "step": 3000
2311
+ },
2312
+ {
2313
+ "epoch": 2.1382751247327154,
2314
+ "eval/nano_beir.climate_fever": 0.2710241560666136,
2315
+ "step": 3000
2316
+ },
2317
+ {
2318
+ "epoch": 2.1382751247327154,
2319
+ "eval/nano_beir.dbpedia_entity": 0.3285043023237046,
2320
+ "step": 3000
2321
+ },
2322
+ {
2323
+ "epoch": 2.1382751247327154,
2324
+ "eval/nano_beir.fever": 0.6331373355635795,
2325
+ "step": 3000
2326
+ },
2327
+ {
2328
+ "epoch": 2.1382751247327154,
2329
+ "eval/nano_beir.fiqa": 0.4150357832110943,
2330
+ "step": 3000
2331
+ },
2332
+ {
2333
+ "epoch": 2.1382751247327154,
2334
+ "eval/nano_beir.hotpotqa": 0.68158322360907,
2335
+ "step": 3000
2336
+ },
2337
+ {
2338
+ "epoch": 2.1382751247327154,
2339
+ "eval/nano_beir.nfcorpus": 0.3531375464357257,
2340
+ "step": 3000
2341
+ },
2342
+ {
2343
+ "epoch": 2.1382751247327154,
2344
+ "eval/nano_beir.nq": 0.41568885251388027,
2345
+ "step": 3000
2346
+ },
2347
+ {
2348
+ "epoch": 2.1382751247327154,
2349
+ "eval/nano_beir.quora": 0.8349377083144377,
2350
+ "step": 3000
2351
+ },
2352
+ {
2353
+ "epoch": 2.1382751247327154,
2354
+ "eval/nano_beir.scidocs": 0.24315533084930935,
2355
+ "step": 3000
2356
+ },
2357
+ {
2358
+ "epoch": 2.1382751247327154,
2359
+ "eval/nano_beir.scifact": 0.7290861351959168,
2360
+ "step": 3000
2361
+ },
2362
+ {
2363
+ "epoch": 2.1382751247327154,
2364
+ "eval/nano_beir.webis_touche2020": 0.28783156872457877,
2365
+ "step": 3000
2366
+ },
2367
+ {
2368
+ "epoch": 2.1382751247327154,
2369
+ "eval/avg": 0.47347706398997325,
2370
+ "step": 3000
2371
+ },
2372
+ {
2373
+ "epoch": 2.1454027084818246,
2374
+ "grad_norm": 5.75,
2375
+ "learning_rate": 4.100837805544279e-05,
2376
+ "loss": 0.7733,
2377
+ "step": 3010
2378
+ },
2379
+ {
2380
+ "epoch": 2.1525302922309337,
2381
+ "grad_norm": 4.28125,
2382
+ "learning_rate": 4.0665224844564053e-05,
2383
+ "loss": 0.7141,
2384
+ "step": 3020
2385
+ },
2386
+ {
2387
+ "epoch": 2.159657875980043,
2388
+ "grad_norm": 6.8125,
2389
+ "learning_rate": 4.032252659745699e-05,
2390
+ "loss": 0.6352,
2391
+ "step": 3030
2392
+ },
2393
+ {
2394
+ "epoch": 2.1667854597291516,
2395
+ "grad_norm": 4.75,
2396
+ "learning_rate": 3.9980300016750694e-05,
2397
+ "loss": 0.8184,
2398
+ "step": 3040
2399
+ },
2400
+ {
2401
+ "epoch": 2.1739130434782608,
2402
+ "grad_norm": 5.0625,
2403
+ "learning_rate": 3.963856178208588e-05,
2404
+ "loss": 0.7785,
2405
+ "step": 3050
2406
+ },
2407
+ {
2408
+ "epoch": 2.18104062722737,
2409
+ "grad_norm": 4.875,
2410
+ "learning_rate": 3.9297328549302e-05,
2411
+ "loss": 0.7565,
2412
+ "step": 3060
2413
+ },
2414
+ {
2415
+ "epoch": 2.188168210976479,
2416
+ "grad_norm": 5.90625,
2417
+ "learning_rate": 3.895661694962542e-05,
2418
+ "loss": 0.7608,
2419
+ "step": 3070
2420
+ },
2421
+ {
2422
+ "epoch": 2.1952957947255882,
2423
+ "grad_norm": 5.90625,
2424
+ "learning_rate": 3.8616443588858805e-05,
2425
+ "loss": 0.7354,
2426
+ "step": 3080
2427
+ },
2428
+ {
2429
+ "epoch": 2.202423378474697,
2430
+ "grad_norm": 5.28125,
2431
+ "learning_rate": 3.827682504657187e-05,
2432
+ "loss": 0.7828,
2433
+ "step": 3090
2434
+ },
2435
+ {
2436
+ "epoch": 2.209550962223806,
2437
+ "grad_norm": 7.1875,
2438
+ "learning_rate": 3.7937777875293244e-05,
2439
+ "loss": 0.7934,
2440
+ "step": 3100
2441
+ },
2442
+ {
2443
+ "epoch": 2.2166785459729152,
2444
+ "grad_norm": 3.859375,
2445
+ "learning_rate": 3.759931859970374e-05,
2446
+ "loss": 0.6331,
2447
+ "step": 3110
2448
+ },
2449
+ {
2450
+ "epoch": 2.2238061297220244,
2451
+ "grad_norm": 5.09375,
2452
+ "learning_rate": 3.72614637158309e-05,
2453
+ "loss": 0.7351,
2454
+ "step": 3120
2455
+ },
2456
+ {
2457
+ "epoch": 2.230933713471133,
2458
+ "grad_norm": 6.15625,
2459
+ "learning_rate": 3.692422969024516e-05,
2460
+ "loss": 0.7856,
2461
+ "step": 3130
2462
+ },
2463
+ {
2464
+ "epoch": 2.2380612972202423,
2465
+ "grad_norm": 5.5,
2466
+ "learning_rate": 3.6587632959257165e-05,
2467
+ "loss": 0.7819,
2468
+ "step": 3140
2469
+ },
2470
+ {
2471
+ "epoch": 2.2451888809693514,
2472
+ "grad_norm": 5.09375,
2473
+ "learning_rate": 3.625168992811671e-05,
2474
+ "loss": 0.7397,
2475
+ "step": 3150
2476
+ },
2477
+ {
2478
+ "epoch": 2.2523164647184606,
2479
+ "grad_norm": 5.21875,
2480
+ "learning_rate": 3.591641697021317e-05,
2481
+ "loss": 0.7588,
2482
+ "step": 3160
2483
+ },
2484
+ {
2485
+ "epoch": 2.2594440484675697,
2486
+ "grad_norm": 5.5625,
2487
+ "learning_rate": 3.558183042627755e-05,
2488
+ "loss": 0.751,
2489
+ "step": 3170
2490
+ },
2491
+ {
2492
+ "epoch": 2.2665716322166785,
2493
+ "grad_norm": 5.46875,
2494
+ "learning_rate": 3.524794660358593e-05,
2495
+ "loss": 0.8684,
2496
+ "step": 3180
2497
+ },
2498
+ {
2499
+ "epoch": 2.2736992159657876,
2500
+ "grad_norm": 4.5625,
2501
+ "learning_rate": 3.491478177516484e-05,
2502
+ "loss": 0.7863,
2503
+ "step": 3190
2504
+ },
2505
+ {
2506
+ "epoch": 2.2808267997148968,
2507
+ "grad_norm": 5.25,
2508
+ "learning_rate": 3.4582352178997935e-05,
2509
+ "loss": 0.7262,
2510
+ "step": 3200
2511
+ },
2512
+ {
2513
+ "epoch": 2.2879543834640055,
2514
+ "grad_norm": 4.0,
2515
+ "learning_rate": 3.425067401723477e-05,
2516
+ "loss": 0.7284,
2517
+ "step": 3210
2518
+ },
2519
+ {
2520
+ "epoch": 2.2950819672131146,
2521
+ "grad_norm": 5.3125,
2522
+ "learning_rate": 3.3919763455401015e-05,
2523
+ "loss": 0.7184,
2524
+ "step": 3220
2525
+ },
2526
+ {
2527
+ "epoch": 2.302209550962224,
2528
+ "grad_norm": 5.21875,
2529
+ "learning_rate": 3.358963662161062e-05,
2530
+ "loss": 0.8039,
2531
+ "step": 3230
2532
+ },
2533
+ {
2534
+ "epoch": 2.309337134711333,
2535
+ "grad_norm": 4.28125,
2536
+ "learning_rate": 3.326030960577972e-05,
2537
+ "loss": 0.7646,
2538
+ "step": 3240
2539
+ },
2540
+ {
2541
+ "epoch": 2.316464718460442,
2542
+ "grad_norm": 5.96875,
2543
+ "learning_rate": 3.293179845884245e-05,
2544
+ "loss": 0.7086,
2545
+ "step": 3250
2546
+ },
2547
+ {
2548
+ "epoch": 2.323592302209551,
2549
+ "grad_norm": 5.6875,
2550
+ "learning_rate": 3.2604119191968654e-05,
2551
+ "loss": 0.8153,
2552
+ "step": 3260
2553
+ },
2554
+ {
2555
+ "epoch": 2.33071988595866,
2556
+ "grad_norm": 6.125,
2557
+ "learning_rate": 3.2277287775783525e-05,
2558
+ "loss": 0.7863,
2559
+ "step": 3270
2560
+ },
2561
+ {
2562
+ "epoch": 2.337847469707769,
2563
+ "grad_norm": 5.90625,
2564
+ "learning_rate": 3.195132013958918e-05,
2565
+ "loss": 0.7569,
2566
+ "step": 3280
2567
+ },
2568
+ {
2569
+ "epoch": 2.3449750534568783,
2570
+ "grad_norm": 4.34375,
2571
+ "learning_rate": 3.162623217058834e-05,
2572
+ "loss": 0.7193,
2573
+ "step": 3290
2574
+ },
2575
+ {
2576
+ "epoch": 2.352102637205987,
2577
+ "grad_norm": 7.15625,
2578
+ "learning_rate": 3.130203971310999e-05,
2579
+ "loss": 0.847,
2580
+ "step": 3300
2581
+ },
2582
+ {
2583
+ "epoch": 2.359230220955096,
2584
+ "grad_norm": 4.6875,
2585
+ "learning_rate": 3.097875856783713e-05,
2586
+ "loss": 0.7884,
2587
+ "step": 3310
2588
+ },
2589
+ {
2590
+ "epoch": 2.3663578047042053,
2591
+ "grad_norm": 4.46875,
2592
+ "learning_rate": 3.065640449103669e-05,
2593
+ "loss": 0.7849,
2594
+ "step": 3320
2595
+ },
2596
+ {
2597
+ "epoch": 2.3734853884533145,
2598
+ "grad_norm": 4.84375,
2599
+ "learning_rate": 3.033499319379163e-05,
2600
+ "loss": 0.7069,
2601
+ "step": 3330
2602
+ },
2603
+ {
2604
+ "epoch": 2.3806129722024236,
2605
+ "grad_norm": 4.28125,
2606
+ "learning_rate": 3.001454034123512e-05,
2607
+ "loss": 0.6801,
2608
+ "step": 3340
2609
+ },
2610
+ {
2611
+ "epoch": 2.3877405559515323,
2612
+ "grad_norm": 5.34375,
2613
+ "learning_rate": 2.9695061551787108e-05,
2614
+ "loss": 0.7755,
2615
+ "step": 3350
2616
+ },
2617
+ {
2618
+ "epoch": 2.3948681397006415,
2619
+ "grad_norm": 4.21875,
2620
+ "learning_rate": 2.9376572396393048e-05,
2621
+ "loss": 0.8031,
2622
+ "step": 3360
2623
+ },
2624
+ {
2625
+ "epoch": 2.4019957234497507,
2626
+ "grad_norm": 5.34375,
2627
+ "learning_rate": 2.9059088397765088e-05,
2628
+ "loss": 0.8501,
2629
+ "step": 3370
2630
+ },
2631
+ {
2632
+ "epoch": 2.4091233071988594,
2633
+ "grad_norm": 4.4375,
2634
+ "learning_rate": 2.874262502962537e-05,
2635
+ "loss": 0.8451,
2636
+ "step": 3380
2637
+ },
2638
+ {
2639
+ "epoch": 2.4162508909479685,
2640
+ "grad_norm": 4.3125,
2641
+ "learning_rate": 2.8427197715952047e-05,
2642
+ "loss": 0.7452,
2643
+ "step": 3390
2644
+ },
2645
+ {
2646
+ "epoch": 2.4233784746970777,
2647
+ "grad_norm": 4.8125,
2648
+ "learning_rate": 2.811282183022736e-05,
2649
+ "loss": 0.6863,
2650
+ "step": 3400
2651
+ },
2652
+ {
2653
+ "epoch": 2.430506058446187,
2654
+ "grad_norm": 5.5,
2655
+ "learning_rate": 2.7799512694688468e-05,
2656
+ "loss": 0.811,
2657
+ "step": 3410
2658
+ },
2659
+ {
2660
+ "epoch": 2.437633642195296,
2661
+ "grad_norm": 4.375,
2662
+ "learning_rate": 2.7487285579580637e-05,
2663
+ "loss": 0.7234,
2664
+ "step": 3420
2665
+ },
2666
+ {
2667
+ "epoch": 2.4447612259444047,
2668
+ "grad_norm": 5.0625,
2669
+ "learning_rate": 2.7176155702412937e-05,
2670
+ "loss": 0.8367,
2671
+ "step": 3430
2672
+ },
2673
+ {
2674
+ "epoch": 2.451888809693514,
2675
+ "grad_norm": 5.90625,
2676
+ "learning_rate": 2.686613822721666e-05,
2677
+ "loss": 0.6122,
2678
+ "step": 3440
2679
+ },
2680
+ {
2681
+ "epoch": 2.459016393442623,
2682
+ "grad_norm": 4.4375,
2683
+ "learning_rate": 2.6557248263806174e-05,
2684
+ "loss": 0.7393,
2685
+ "step": 3450
2686
+ },
2687
+ {
2688
+ "epoch": 2.466143977191732,
2689
+ "grad_norm": 5.0,
2690
+ "learning_rate": 2.6249500867042522e-05,
2691
+ "loss": 0.7783,
2692
+ "step": 3460
2693
+ },
2694
+ {
2695
+ "epoch": 2.473271560940841,
2696
+ "grad_norm": 5.1875,
2697
+ "learning_rate": 2.5942911036099655e-05,
2698
+ "loss": 0.6654,
2699
+ "step": 3470
2700
+ },
2701
+ {
2702
+ "epoch": 2.48039914468995,
2703
+ "grad_norm": 4.53125,
2704
+ "learning_rate": 2.5637493713733374e-05,
2705
+ "loss": 0.752,
2706
+ "step": 3480
2707
+ },
2708
+ {
2709
+ "epoch": 2.487526728439059,
2710
+ "grad_norm": 5.1875,
2711
+ "learning_rate": 2.533326378555314e-05,
2712
+ "loss": 0.7761,
2713
+ "step": 3490
2714
+ },
2715
+ {
2716
+ "epoch": 2.4946543121881684,
2717
+ "grad_norm": 5.53125,
2718
+ "learning_rate": 2.5030236079296444e-05,
2719
+ "loss": 0.7594,
2720
+ "step": 3500
2721
+ },
2722
+ {
2723
+ "epoch": 2.5017818959372775,
2724
+ "grad_norm": 5.6875,
2725
+ "learning_rate": 2.4728425364106135e-05,
2726
+ "loss": 0.7519,
2727
+ "step": 3510
2728
+ },
2729
+ {
2730
+ "epoch": 2.5089094796863862,
2731
+ "grad_norm": 5.1875,
2732
+ "learning_rate": 2.442784634981071e-05,
2733
+ "loss": 0.7527,
2734
+ "step": 3520
2735
+ },
2736
+ {
2737
+ "epoch": 2.5160370634354954,
2738
+ "grad_norm": 4.28125,
2739
+ "learning_rate": 2.412851368620726e-05,
2740
+ "loss": 0.6595,
2741
+ "step": 3530
2742
+ },
2743
+ {
2744
+ "epoch": 2.5231646471846045,
2745
+ "grad_norm": 4.53125,
2746
+ "learning_rate": 2.383044196234753e-05,
2747
+ "loss": 0.7621,
2748
+ "step": 3540
2749
+ },
2750
+ {
2751
+ "epoch": 2.5302922309337132,
2752
+ "grad_norm": 4.28125,
2753
+ "learning_rate": 2.3533645705826806e-05,
2754
+ "loss": 0.7766,
2755
+ "step": 3550
2756
+ },
2757
+ {
2758
+ "epoch": 2.5374198146828224,
2759
+ "grad_norm": 4.625,
2760
+ "learning_rate": 2.323813938207593e-05,
2761
+ "loss": 0.7717,
2762
+ "step": 3560
2763
+ },
2764
+ {
2765
+ "epoch": 2.5445473984319316,
2766
+ "grad_norm": 5.625,
2767
+ "learning_rate": 2.294393739365621e-05,
2768
+ "loss": 0.6604,
2769
+ "step": 3570
2770
+ },
2771
+ {
2772
+ "epoch": 2.5516749821810407,
2773
+ "grad_norm": 4.03125,
2774
+ "learning_rate": 2.265105407955752e-05,
2775
+ "loss": 0.7614,
2776
+ "step": 3580
2777
+ },
2778
+ {
2779
+ "epoch": 2.55880256593015,
2780
+ "grad_norm": 5.59375,
2781
+ "learning_rate": 2.235950371449938e-05,
2782
+ "loss": 0.7702,
2783
+ "step": 3590
2784
+ },
2785
+ {
2786
+ "epoch": 2.5659301496792586,
2787
+ "grad_norm": 5.96875,
2788
+ "learning_rate": 2.2069300508235275e-05,
2789
+ "loss": 0.8866,
2790
+ "step": 3600
2791
+ },
2792
+ {
2793
+ "epoch": 2.5730577334283677,
2794
+ "grad_norm": 6.21875,
2795
+ "learning_rate": 2.1780458604860056e-05,
2796
+ "loss": 0.8359,
2797
+ "step": 3610
2798
+ },
2799
+ {
2800
+ "epoch": 2.580185317177477,
2801
+ "grad_norm": 4.75,
2802
+ "learning_rate": 2.14929920821206e-05,
2803
+ "loss": 0.7036,
2804
+ "step": 3620
2805
+ },
2806
+ {
2807
+ "epoch": 2.5873129009265856,
2808
+ "grad_norm": 6.25,
2809
+ "learning_rate": 2.1206914950729672e-05,
2810
+ "loss": 0.7946,
2811
+ "step": 3630
2812
+ },
2813
+ {
2814
+ "epoch": 2.5944404846756948,
2815
+ "grad_norm": 8.1875,
2816
+ "learning_rate": 2.0922241153683066e-05,
2817
+ "loss": 0.7035,
2818
+ "step": 3640
2819
+ },
2820
+ {
2821
+ "epoch": 2.601568068424804,
2822
+ "grad_norm": 5.0625,
2823
+ "learning_rate": 2.0638984565580023e-05,
2824
+ "loss": 0.7456,
2825
+ "step": 3650
2826
+ },
2827
+ {
2828
+ "epoch": 2.608695652173913,
2829
+ "grad_norm": 5.6875,
2830
+ "learning_rate": 2.035715899194704e-05,
2831
+ "loss": 0.7462,
2832
+ "step": 3660
2833
+ },
2834
+ {
2835
+ "epoch": 2.6158232359230222,
2836
+ "grad_norm": 4.65625,
2837
+ "learning_rate": 2.007677816856498e-05,
2838
+ "loss": 0.7088,
2839
+ "step": 3670
2840
+ },
2841
+ {
2842
+ "epoch": 2.6229508196721314,
2843
+ "grad_norm": 4.84375,
2844
+ "learning_rate": 1.9797855760799612e-05,
2845
+ "loss": 0.721,
2846
+ "step": 3680
2847
+ },
2848
+ {
2849
+ "epoch": 2.63007840342124,
2850
+ "grad_norm": 5.84375,
2851
+ "learning_rate": 1.9520405362935594e-05,
2852
+ "loss": 0.7292,
2853
+ "step": 3690
2854
+ },
2855
+ {
2856
+ "epoch": 2.6372059871703493,
2857
+ "grad_norm": 5.0,
2858
+ "learning_rate": 1.9244440497513893e-05,
2859
+ "loss": 0.7031,
2860
+ "step": 3700
2861
+ },
2862
+ {
2863
+ "epoch": 2.6443335709194584,
2864
+ "grad_norm": 4.3125,
2865
+ "learning_rate": 1.896997461467272e-05,
2866
+ "loss": 0.696,
2867
+ "step": 3710
2868
+ },
2869
+ {
2870
+ "epoch": 2.651461154668567,
2871
+ "grad_norm": 4.875,
2872
+ "learning_rate": 1.869702109149199e-05,
2873
+ "loss": 0.749,
2874
+ "step": 3720
2875
+ },
2876
+ {
2877
+ "epoch": 2.6585887384176763,
2878
+ "grad_norm": 6.21875,
2879
+ "learning_rate": 1.842559323134136e-05,
2880
+ "loss": 0.7282,
2881
+ "step": 3730
2882
+ },
2883
+ {
2884
+ "epoch": 2.6657163221667854,
2885
+ "grad_norm": 5.0625,
2886
+ "learning_rate": 1.8155704263231776e-05,
2887
+ "loss": 0.6514,
2888
+ "step": 3740
2889
+ },
2890
+ {
2891
+ "epoch": 2.6728439059158946,
2892
+ "grad_norm": 4.9375,
2893
+ "learning_rate": 1.788736734117078e-05,
2894
+ "loss": 0.7496,
2895
+ "step": 3750
2896
+ },
2897
+ {
2898
+ "epoch": 2.6799714896650038,
2899
+ "grad_norm": 4.5,
2900
+ "learning_rate": 1.762059554352143e-05,
2901
+ "loss": 0.8261,
2902
+ "step": 3760
2903
+ },
2904
+ {
2905
+ "epoch": 2.6870990734141125,
2906
+ "grad_norm": 4.8125,
2907
+ "learning_rate": 1.7355401872364758e-05,
2908
+ "loss": 0.8161,
2909
+ "step": 3770
2910
+ },
2911
+ {
2912
+ "epoch": 2.6942266571632216,
2913
+ "grad_norm": 5.125,
2914
+ "learning_rate": 1.709179925286617e-05,
2915
+ "loss": 0.682,
2916
+ "step": 3780
2917
+ },
2918
+ {
2919
+ "epoch": 2.701354240912331,
2920
+ "grad_norm": 4.90625,
2921
+ "learning_rate": 1.6829800532645447e-05,
2922
+ "loss": 0.7542,
2923
+ "step": 3790
2924
+ },
2925
+ {
2926
+ "epoch": 2.7084818246614395,
2927
+ "grad_norm": 6.46875,
2928
+ "learning_rate": 1.6569418481150595e-05,
2929
+ "loss": 0.8126,
2930
+ "step": 3800
2931
+ },
2932
+ {
2933
+ "epoch": 2.7156094084105487,
2934
+ "grad_norm": 4.6875,
2935
+ "learning_rate": 1.6310665789035467e-05,
2936
+ "loss": 0.7445,
2937
+ "step": 3810
2938
+ },
2939
+ {
2940
+ "epoch": 2.722736992159658,
2941
+ "grad_norm": 4.9375,
2942
+ "learning_rate": 1.605355506754121e-05,
2943
+ "loss": 0.8296,
2944
+ "step": 3820
2945
+ },
2946
+ {
2947
+ "epoch": 2.729864575908767,
2948
+ "grad_norm": 4.84375,
2949
+ "learning_rate": 1.5798098847881666e-05,
2950
+ "loss": 0.6684,
2951
+ "step": 3830
2952
+ },
2953
+ {
2954
+ "epoch": 2.736992159657876,
2955
+ "grad_norm": 4.5625,
2956
+ "learning_rate": 1.554430958063259e-05,
2957
+ "loss": 0.7063,
2958
+ "step": 3840
2959
+ },
2960
+ {
2961
+ "epoch": 2.7441197434069853,
2962
+ "grad_norm": 4.625,
2963
+ "learning_rate": 1.529219963512481e-05,
2964
+ "loss": 0.6668,
2965
+ "step": 3850
2966
+ },
2967
+ {
2968
+ "epoch": 2.751247327156094,
2969
+ "grad_norm": 5.21875,
2970
+ "learning_rate": 1.5041781298841423e-05,
2971
+ "loss": 0.7464,
2972
+ "step": 3860
2973
+ },
2974
+ {
2975
+ "epoch": 2.758374910905203,
2976
+ "grad_norm": 4.78125,
2977
+ "learning_rate": 1.4793066776818842e-05,
2978
+ "loss": 0.7625,
2979
+ "step": 3870
2980
+ },
2981
+ {
2982
+ "epoch": 2.7655024946543123,
2983
+ "grad_norm": 5.625,
2984
+ "learning_rate": 1.4546068191051987e-05,
2985
+ "loss": 0.738,
2986
+ "step": 3880
2987
+ },
2988
+ {
2989
+ "epoch": 2.772630078403421,
2990
+ "grad_norm": 6.125,
2991
+ "learning_rate": 1.4300797579903474e-05,
2992
+ "loss": 0.8185,
2993
+ "step": 3890
2994
+ },
2995
+ {
2996
+ "epoch": 2.77975766215253,
2997
+ "grad_norm": 5.4375,
2998
+ "learning_rate": 1.4057266897516841e-05,
2999
+ "loss": 0.7021,
3000
+ "step": 3900
3001
+ },
3002
+ {
3003
+ "epoch": 2.7868852459016393,
3004
+ "grad_norm": 6.78125,
3005
+ "learning_rate": 1.3815488013233984e-05,
3006
+ "loss": 0.8398,
3007
+ "step": 3910
3008
+ },
3009
+ {
3010
+ "epoch": 2.7940128296507485,
3011
+ "grad_norm": 6.03125,
3012
+ "learning_rate": 1.3575472711016635e-05,
3013
+ "loss": 0.8821,
3014
+ "step": 3920
3015
+ },
3016
+ {
3017
+ "epoch": 2.8011404133998576,
3018
+ "grad_norm": 4.75,
3019
+ "learning_rate": 1.3337232688872009e-05,
3020
+ "loss": 0.7238,
3021
+ "step": 3930
3022
+ },
3023
+ {
3024
+ "epoch": 2.8082679971489664,
3025
+ "grad_norm": 5.1875,
3026
+ "learning_rate": 1.3100779558282672e-05,
3027
+ "loss": 0.6887,
3028
+ "step": 3940
3029
+ },
3030
+ {
3031
+ "epoch": 2.8153955808980755,
3032
+ "grad_norm": 6.03125,
3033
+ "learning_rate": 1.2866124843640615e-05,
3034
+ "loss": 0.7804,
3035
+ "step": 3950
3036
+ },
3037
+ {
3038
+ "epoch": 2.8225231646471847,
3039
+ "grad_norm": 5.78125,
3040
+ "learning_rate": 1.2633279981685608e-05,
3041
+ "loss": 0.871,
3042
+ "step": 3960
3043
+ },
3044
+ {
3045
+ "epoch": 2.8296507483962934,
3046
+ "grad_norm": 4.40625,
3047
+ "learning_rate": 1.240225632094773e-05,
3048
+ "loss": 0.764,
3049
+ "step": 3970
3050
+ },
3051
+ {
3052
+ "epoch": 2.8367783321454025,
3053
+ "grad_norm": 4.65625,
3054
+ "learning_rate": 1.217306512119425e-05,
3055
+ "loss": 0.7754,
3056
+ "step": 3980
3057
+ },
3058
+ {
3059
+ "epoch": 2.8439059158945117,
3060
+ "grad_norm": 5.71875,
3061
+ "learning_rate": 1.194571755288092e-05,
3062
+ "loss": 0.7185,
3063
+ "step": 3990
3064
+ },
3065
+ {
3066
+ "epoch": 2.851033499643621,
3067
+ "grad_norm": 4.9375,
3068
+ "learning_rate": 1.1720224696607474e-05,
3069
+ "loss": 0.6823,
3070
+ "step": 4000
3071
+ },
3072
+ {
3073
+ "epoch": 2.851033499643621,
3074
+ "eval/nano_beir.arguana": 0.47570368623108616,
3075
+ "step": 4000
3076
+ },
3077
+ {
3078
+ "epoch": 2.851033499643621,
3079
+ "eval/nano_beir.climate_fever": 0.27306481927622206,
3080
+ "step": 4000
3081
+ },
3082
+ {
3083
+ "epoch": 2.851033499643621,
3084
+ "eval/nano_beir.dbpedia_entity": 0.3271142777426427,
3085
+ "step": 4000
3086
+ },
3087
+ {
3088
+ "epoch": 2.851033499643621,
3089
+ "eval/nano_beir.fever": 0.6267850303425887,
3090
+ "step": 4000
3091
+ },
3092
+ {
3093
+ "epoch": 2.851033499643621,
3094
+ "eval/nano_beir.fiqa": 0.42142137153910825,
3095
+ "step": 4000
3096
+ },
3097
+ {
3098
+ "epoch": 2.851033499643621,
3099
+ "eval/nano_beir.hotpotqa": 0.6860936998599775,
3100
+ "step": 4000
3101
+ },
3102
+ {
3103
+ "epoch": 2.851033499643621,
3104
+ "eval/nano_beir.nfcorpus": 0.3525089997861853,
3105
+ "step": 4000
3106
+ },
3107
+ {
3108
+ "epoch": 2.851033499643621,
3109
+ "eval/nano_beir.nq": 0.40835064000757015,
3110
+ "step": 4000
3111
+ },
3112
+ {
3113
+ "epoch": 2.851033499643621,
3114
+ "eval/nano_beir.quora": 0.8347230240552653,
3115
+ "step": 4000
3116
+ },
3117
+ {
3118
+ "epoch": 2.851033499643621,
3119
+ "eval/nano_beir.scidocs": 0.2466296038296945,
3120
+ "step": 4000
3121
+ },
3122
+ {
3123
+ "epoch": 2.851033499643621,
3124
+ "eval/nano_beir.scifact": 0.7298579737170866,
3125
+ "step": 4000
3126
+ },
3127
+ {
3128
+ "epoch": 2.851033499643621,
3129
+ "eval/nano_beir.webis_touche2020": 0.2863659661103585,
3130
+ "step": 4000
3131
+ },
3132
+ {
3133
+ "epoch": 2.851033499643621,
3134
+ "eval/avg": 0.4723849243748155,
3135
+ "step": 4000
3136
+ },
3137
+ {
3138
+ "epoch": 2.85816108339273,
3139
+ "grad_norm": 5.3125,
3140
+ "learning_rate": 1.1496597542577603e-05,
3141
+ "loss": 0.8187,
3142
+ "step": 4010
3143
+ },
3144
+ {
3145
+ "epoch": 2.865288667141839,
3146
+ "grad_norm": 4.40625,
3147
+ "learning_rate": 1.1274846990063315e-05,
3148
+ "loss": 0.7384,
3149
+ "step": 4020
3150
+ },
3151
+ {
3152
+ "epoch": 2.872416250890948,
3153
+ "grad_norm": 5.375,
3154
+ "learning_rate": 1.1054983846873685e-05,
3155
+ "loss": 0.7379,
3156
+ "step": 4030
3157
+ },
3158
+ {
3159
+ "epoch": 2.879543834640057,
3160
+ "grad_norm": 6.5,
3161
+ "learning_rate": 1.0837018828828133e-05,
3162
+ "loss": 0.8139,
3163
+ "step": 4040
3164
+ },
3165
+ {
3166
+ "epoch": 2.886671418389166,
3167
+ "grad_norm": 5.96875,
3168
+ "learning_rate": 1.0620962559234143e-05,
3169
+ "loss": 0.8204,
3170
+ "step": 4050
3171
+ },
3172
+ {
3173
+ "epoch": 2.893799002138275,
3174
+ "grad_norm": 5.8125,
3175
+ "learning_rate": 1.0406825568369477e-05,
3176
+ "loss": 0.8589,
3177
+ "step": 4060
3178
+ },
3179
+ {
3180
+ "epoch": 2.900926585887384,
3181
+ "grad_norm": 4.96875,
3182
+ "learning_rate": 1.0194618292968972e-05,
3183
+ "loss": 0.7727,
3184
+ "step": 4070
3185
+ },
3186
+ {
3187
+ "epoch": 2.908054169636493,
3188
+ "grad_norm": 5.0625,
3189
+ "learning_rate": 9.984351075715848e-06,
3190
+ "loss": 0.7422,
3191
+ "step": 4080
3192
+ },
3193
+ {
3194
+ "epoch": 2.9151817533856024,
3195
+ "grad_norm": 4.8125,
3196
+ "learning_rate": 9.77603416473763e-06,
3197
+ "loss": 0.8394,
3198
+ "step": 4090
3199
+ },
3200
+ {
3201
+ "epoch": 2.9223093371347115,
3202
+ "grad_norm": 5.53125,
3203
+ "learning_rate": 9.569677713106674e-06,
3204
+ "loss": 0.6983,
3205
+ "step": 4100
3206
+ },
3207
+ {
3208
+ "epoch": 2.9294369208838202,
3209
+ "grad_norm": 5.9375,
3210
+ "learning_rate": 9.365291778345303e-06,
3211
+ "loss": 0.7259,
3212
+ "step": 4110
3213
+ },
3214
+ {
3215
+ "epoch": 2.9365645046329294,
3216
+ "grad_norm": 5.15625,
3217
+ "learning_rate": 9.162886321935632e-06,
3218
+ "loss": 0.7249,
3219
+ "step": 4120
3220
+ },
3221
+ {
3222
+ "epoch": 2.9436920883820386,
3223
+ "grad_norm": 5.03125,
3224
+ "learning_rate": 8.962471208834055e-06,
3225
+ "loss": 0.7308,
3226
+ "step": 4130
3227
+ },
3228
+ {
3229
+ "epoch": 2.9508196721311473,
3230
+ "grad_norm": 4.90625,
3231
+ "learning_rate": 8.764056206990445e-06,
3232
+ "loss": 0.6875,
3233
+ "step": 4140
3234
+ },
3235
+ {
3236
+ "epoch": 2.9579472558802564,
3237
+ "grad_norm": 5.40625,
3238
+ "learning_rate": 8.567650986872062e-06,
3239
+ "loss": 0.7786,
3240
+ "step": 4150
3241
+ },
3242
+ {
3243
+ "epoch": 2.9650748396293656,
3244
+ "grad_norm": 4.46875,
3245
+ "learning_rate": 8.373265120992252e-06,
3246
+ "loss": 0.7357,
3247
+ "step": 4160
3248
+ },
3249
+ {
3250
+ "epoch": 2.9722024233784747,
3251
+ "grad_norm": 5.46875,
3252
+ "learning_rate": 8.180908083443884e-06,
3253
+ "loss": 0.7817,
3254
+ "step": 4170
3255
+ },
3256
+ {
3257
+ "epoch": 2.979330007127584,
3258
+ "grad_norm": 4.75,
3259
+ "learning_rate": 7.99058924943759e-06,
3260
+ "loss": 0.7892,
3261
+ "step": 4180
3262
+ },
3263
+ {
3264
+ "epoch": 2.986457590876693,
3265
+ "grad_norm": 6.25,
3266
+ "learning_rate": 7.802317894844835e-06,
3267
+ "loss": 0.7669,
3268
+ "step": 4190
3269
+ },
3270
+ {
3271
+ "epoch": 2.9935851746258018,
3272
+ "grad_norm": 5.5,
3273
+ "learning_rate": 7.6161031957458494e-06,
3274
+ "loss": 0.8586,
3275
+ "step": 4200
3276
+ },
3277
+ {
3278
+ "epoch": 3.000712758374911,
3279
+ "grad_norm": 4.40625,
3280
+ "learning_rate": 7.43195422798233e-06,
3281
+ "loss": 0.7776,
3282
+ "step": 4210
3283
+ },
3284
+ {
3285
+ "epoch": 3.00784034212402,
3286
+ "grad_norm": 5.34375,
3287
+ "learning_rate": 7.249879966715173e-06,
3288
+ "loss": 0.7948,
3289
+ "step": 4220
3290
+ },
3291
+ {
3292
+ "epoch": 3.0149679258731292,
3293
+ "grad_norm": 5.6875,
3294
+ "learning_rate": 7.0698892859870245e-06,
3295
+ "loss": 0.7505,
3296
+ "step": 4230
3297
+ },
3298
+ {
3299
+ "epoch": 3.022095509622238,
3300
+ "grad_norm": 5.5,
3301
+ "learning_rate": 6.891990958289723e-06,
3302
+ "loss": 0.8047,
3303
+ "step": 4240
3304
+ },
3305
+ {
3306
+ "epoch": 3.029223093371347,
3307
+ "grad_norm": 5.1875,
3308
+ "learning_rate": 6.716193654136788e-06,
3309
+ "loss": 0.7658,
3310
+ "step": 4250
3311
+ },
3312
+ {
3313
+ "epoch": 3.0363506771204563,
3314
+ "grad_norm": 4.8125,
3315
+ "learning_rate": 6.542505941640803e-06,
3316
+ "loss": 0.675,
3317
+ "step": 4260
3318
+ },
3319
+ {
3320
+ "epoch": 3.0434782608695654,
3321
+ "grad_norm": 5.09375,
3322
+ "learning_rate": 6.370936286095841e-06,
3323
+ "loss": 0.8296,
3324
+ "step": 4270
3325
+ },
3326
+ {
3327
+ "epoch": 3.050605844618674,
3328
+ "grad_norm": 4.875,
3329
+ "learning_rate": 6.201493049564882e-06,
3330
+ "loss": 0.7138,
3331
+ "step": 4280
3332
+ },
3333
+ {
3334
+ "epoch": 3.0577334283677833,
3335
+ "grad_norm": 3.984375,
3336
+ "learning_rate": 6.034184490472195e-06,
3337
+ "loss": 0.6907,
3338
+ "step": 4290
3339
+ },
3340
+ {
3341
+ "epoch": 3.0648610121168924,
3342
+ "grad_norm": 5.625,
3343
+ "learning_rate": 5.8690187632009285e-06,
3344
+ "loss": 0.779,
3345
+ "step": 4300
3346
+ },
3347
+ {
3348
+ "epoch": 3.0719885958660016,
3349
+ "grad_norm": 4.5,
3350
+ "learning_rate": 5.706003917695618e-06,
3351
+ "loss": 0.7086,
3352
+ "step": 4310
3353
+ },
3354
+ {
3355
+ "epoch": 3.0791161796151103,
3356
+ "grad_norm": 5.96875,
3357
+ "learning_rate": 5.545147899069836e-06,
3358
+ "loss": 0.7896,
3359
+ "step": 4320
3360
+ },
3361
+ {
3362
+ "epoch": 3.0862437633642195,
3363
+ "grad_norm": 6.5625,
3364
+ "learning_rate": 5.3864585472190255e-06,
3365
+ "loss": 0.7141,
3366
+ "step": 4330
3367
+ },
3368
+ {
3369
+ "epoch": 3.0933713471133286,
3370
+ "grad_norm": 5.78125,
3371
+ "learning_rate": 5.229943596438297e-06,
3372
+ "loss": 0.7517,
3373
+ "step": 4340
3374
+ },
3375
+ {
3376
+ "epoch": 3.1004989308624378,
3377
+ "grad_norm": 5.78125,
3378
+ "learning_rate": 5.075610675045567e-06,
3379
+ "loss": 0.7389,
3380
+ "step": 4350
3381
+ },
3382
+ {
3383
+ "epoch": 3.1076265146115465,
3384
+ "grad_norm": 4.6875,
3385
+ "learning_rate": 4.92346730500966e-06,
3386
+ "loss": 0.6878,
3387
+ "step": 4360
3388
+ },
3389
+ {
3390
+ "epoch": 3.1147540983606556,
3391
+ "grad_norm": 5.96875,
3392
+ "learning_rate": 4.7735209015838e-06,
3393
+ "loss": 0.7404,
3394
+ "step": 4370
3395
+ },
3396
+ {
3397
+ "epoch": 3.121881682109765,
3398
+ "grad_norm": 4.9375,
3399
+ "learning_rate": 4.625778772944156e-06,
3400
+ "loss": 0.702,
3401
+ "step": 4380
3402
+ },
3403
+ {
3404
+ "epoch": 3.129009265858874,
3405
+ "grad_norm": 5.0625,
3406
+ "learning_rate": 4.480248119833641e-06,
3407
+ "loss": 0.7602,
3408
+ "step": 4390
3409
+ },
3410
+ {
3411
+ "epoch": 3.1361368496079827,
3412
+ "grad_norm": 6.5,
3413
+ "learning_rate": 4.33693603521097e-06,
3414
+ "loss": 0.7341,
3415
+ "step": 4400
3416
+ },
3417
+ {
3418
+ "epoch": 3.143264433357092,
3419
+ "grad_norm": 5.75,
3420
+ "learning_rate": 4.195849503904975e-06,
3421
+ "loss": 0.7399,
3422
+ "step": 4410
3423
+ },
3424
+ {
3425
+ "epoch": 3.150392017106201,
3426
+ "grad_norm": 4.1875,
3427
+ "learning_rate": 4.056995402274122e-06,
3428
+ "loss": 0.7524,
3429
+ "step": 4420
3430
+ },
3431
+ {
3432
+ "epoch": 3.15751960085531,
3433
+ "grad_norm": 5.0625,
3434
+ "learning_rate": 3.920380497871473e-06,
3435
+ "loss": 0.78,
3436
+ "step": 4430
3437
+ },
3438
+ {
3439
+ "epoch": 3.1646471846044193,
3440
+ "grad_norm": 5.71875,
3441
+ "learning_rate": 3.7860114491147016e-06,
3442
+ "loss": 0.7816,
3443
+ "step": 4440
3444
+ },
3445
+ {
3446
+ "epoch": 3.171774768353528,
3447
+ "grad_norm": 3.953125,
3448
+ "learning_rate": 3.6538948049616882e-06,
3449
+ "loss": 0.7733,
3450
+ "step": 4450
3451
+ },
3452
+ {
3453
+ "epoch": 3.178902352102637,
3454
+ "grad_norm": 5.4375,
3455
+ "learning_rate": 3.524037004591274e-06,
3456
+ "loss": 0.8646,
3457
+ "step": 4460
3458
+ },
3459
+ {
3460
+ "epoch": 3.1860299358517463,
3461
+ "grad_norm": 4.5,
3462
+ "learning_rate": 3.3964443770894528e-06,
3463
+ "loss": 0.806,
3464
+ "step": 4470
3465
+ },
3466
+ {
3467
+ "epoch": 3.1931575196008555,
3468
+ "grad_norm": 5.25,
3469
+ "learning_rate": 3.271123141140886e-06,
3470
+ "loss": 0.7918,
3471
+ "step": 4480
3472
+ },
3473
+ {
3474
+ "epoch": 3.200285103349964,
3475
+ "grad_norm": 6.0,
3476
+ "learning_rate": 3.148079404725801e-06,
3477
+ "loss": 0.8287,
3478
+ "step": 4490
3479
+ },
3480
+ {
3481
+ "epoch": 3.2074126870990733,
3482
+ "grad_norm": 4.0,
3483
+ "learning_rate": 3.0273191648223287e-06,
3484
+ "loss": 0.6243,
3485
+ "step": 4500
3486
+ },
3487
+ {
3488
+ "epoch": 3.2145402708481825,
3489
+ "grad_norm": 4.9375,
3490
+ "learning_rate": 2.9088483071141983e-06,
3491
+ "loss": 0.7122,
3492
+ "step": 4510
3493
+ },
3494
+ {
3495
+ "epoch": 3.2216678545972917,
3496
+ "grad_norm": 5.1875,
3497
+ "learning_rate": 2.7926726057038666e-06,
3498
+ "loss": 0.7312,
3499
+ "step": 4520
3500
+ },
3501
+ {
3502
+ "epoch": 3.2287954383464004,
3503
+ "grad_norm": 4.5625,
3504
+ "learning_rate": 2.6787977228311334e-06,
3505
+ "loss": 0.7337,
3506
+ "step": 4530
3507
+ },
3508
+ {
3509
+ "epoch": 3.2359230220955095,
3510
+ "grad_norm": 5.28125,
3511
+ "learning_rate": 2.567229208597127e-06,
3512
+ "loss": 0.7347,
3513
+ "step": 4540
3514
+ },
3515
+ {
3516
+ "epoch": 3.2430506058446187,
3517
+ "grad_norm": 4.28125,
3518
+ "learning_rate": 2.4579725006938337e-06,
3519
+ "loss": 0.7344,
3520
+ "step": 4550
3521
+ },
3522
+ {
3523
+ "epoch": 3.250178189593728,
3524
+ "grad_norm": 5.46875,
3525
+ "learning_rate": 2.351032924139063e-06,
3526
+ "loss": 0.8731,
3527
+ "step": 4560
3528
+ },
3529
+ {
3530
+ "epoch": 3.2573057733428366,
3531
+ "grad_norm": 5.15625,
3532
+ "learning_rate": 2.2464156910168954e-06,
3533
+ "loss": 0.7059,
3534
+ "step": 4570
3535
+ },
3536
+ {
3537
+ "epoch": 3.2644333570919457,
3538
+ "grad_norm": 4.875,
3539
+ "learning_rate": 2.1441259002236924e-06,
3540
+ "loss": 0.7404,
3541
+ "step": 4580
3542
+ },
3543
+ {
3544
+ "epoch": 3.271560940841055,
3545
+ "grad_norm": 3.9375,
3546
+ "learning_rate": 2.0441685372195484e-06,
3547
+ "loss": 0.709,
3548
+ "step": 4590
3549
+ },
3550
+ {
3551
+ "epoch": 3.278688524590164,
3552
+ "grad_norm": 4.625,
3553
+ "learning_rate": 1.946548473785309e-06,
3554
+ "loss": 0.732,
3555
+ "step": 4600
3556
+ },
3557
+ {
3558
+ "epoch": 3.285816108339273,
3559
+ "grad_norm": 6.3125,
3560
+ "learning_rate": 1.8512704677851488e-06,
3561
+ "loss": 0.8489,
3562
+ "step": 4610
3563
+ },
3564
+ {
3565
+ "epoch": 3.292943692088382,
3566
+ "grad_norm": 4.5,
3567
+ "learning_rate": 1.758339162934658e-06,
3568
+ "loss": 0.7027,
3569
+ "step": 4620
3570
+ },
3571
+ {
3572
+ "epoch": 3.300071275837491,
3573
+ "grad_norm": 4.59375,
3574
+ "learning_rate": 1.6677590885745386e-06,
3575
+ "loss": 0.7346,
3576
+ "step": 4630
3577
+ },
3578
+ {
3579
+ "epoch": 3.3071988595866,
3580
+ "grad_norm": 4.9375,
3581
+ "learning_rate": 1.5795346594498161e-06,
3582
+ "loss": 0.8141,
3583
+ "step": 4640
3584
+ },
3585
+ {
3586
+ "epoch": 3.3143264433357094,
3587
+ "grad_norm": 4.40625,
3588
+ "learning_rate": 1.4936701754947101e-06,
3589
+ "loss": 0.695,
3590
+ "step": 4650
3591
+ },
3592
+ {
3593
+ "epoch": 3.321454027084818,
3594
+ "grad_norm": 4.6875,
3595
+ "learning_rate": 1.4101698216230253e-06,
3596
+ "loss": 0.7755,
3597
+ "step": 4660
3598
+ },
3599
+ {
3600
+ "epoch": 3.3285816108339272,
3601
+ "grad_norm": 3.953125,
3602
+ "learning_rate": 1.3290376675242023e-06,
3603
+ "loss": 0.7076,
3604
+ "step": 4670
3605
+ },
3606
+ {
3607
+ "epoch": 3.3357091945830364,
3608
+ "grad_norm": 5.59375,
3609
+ "learning_rate": 1.2502776674649774e-06,
3610
+ "loss": 0.8417,
3611
+ "step": 4680
3612
+ },
3613
+ {
3614
+ "epoch": 3.3428367783321455,
3615
+ "grad_norm": 5.625,
3616
+ "learning_rate": 1.1738936600966367e-06,
3617
+ "loss": 0.8167,
3618
+ "step": 4690
3619
+ },
3620
+ {
3621
+ "epoch": 3.3499643620812547,
3622
+ "grad_norm": 5.65625,
3623
+ "learning_rate": 1.0998893682679479e-06,
3624
+ "loss": 0.8371,
3625
+ "step": 4700
3626
+ },
3627
+ {
3628
+ "epoch": 3.3570919458303634,
3629
+ "grad_norm": 5.3125,
3630
+ "learning_rate": 1.0282683988436792e-06,
3631
+ "loss": 0.8034,
3632
+ "step": 4710
3633
+ },
3634
+ {
3635
+ "epoch": 3.3642195295794726,
3636
+ "grad_norm": 6.40625,
3637
+ "learning_rate": 9.590342425288446e-07,
3638
+ "loss": 0.8418,
3639
+ "step": 4720
3640
+ },
3641
+ {
3642
+ "epoch": 3.3713471133285817,
3643
+ "grad_norm": 5.3125,
3644
+ "learning_rate": 8.921902736985399e-07,
3645
+ "loss": 0.7213,
3646
+ "step": 4730
3647
+ },
3648
+ {
3649
+ "epoch": 3.3784746970776904,
3650
+ "grad_norm": 5.8125,
3651
+ "learning_rate": 8.277397502335194e-07,
3652
+ "loss": 0.707,
3653
+ "step": 4740
3654
+ },
3655
+ {
3656
+ "epoch": 3.3856022808267996,
3657
+ "grad_norm": 5.40625,
3658
+ "learning_rate": 7.656858133613498e-07,
3659
+ "loss": 0.8238,
3660
+ "step": 4750
3661
+ },
3662
+ {
3663
+ "epoch": 3.3927298645759087,
3664
+ "grad_norm": 5.1875,
3665
+ "learning_rate": 7.060314875033835e-07,
3666
+ "loss": 0.6872,
3667
+ "step": 4760
3668
+ },
3669
+ {
3670
+ "epoch": 3.399857448325018,
3671
+ "grad_norm": 5.75,
3672
+ "learning_rate": 6.487796801272983e-07,
3673
+ "loss": 0.8421,
3674
+ "step": 4770
3675
+ },
3676
+ {
3677
+ "epoch": 3.406985032074127,
3678
+ "grad_norm": 4.96875,
3679
+ "learning_rate": 5.939331816054161e-07,
3680
+ "loss": 0.784,
3681
+ "step": 4780
3682
+ },
3683
+ {
3684
+ "epoch": 3.4141126158232358,
3685
+ "grad_norm": 4.75,
3686
+ "learning_rate": 5.414946650786957e-07,
3687
+ "loss": 0.728,
3688
+ "step": 4790
3689
+ },
3690
+ {
3691
+ "epoch": 3.421240199572345,
3692
+ "grad_norm": 5.34375,
3693
+ "learning_rate": 4.914666863264528e-07,
3694
+ "loss": 0.8101,
3695
+ "step": 4800
3696
+ },
3697
+ {
3698
+ "epoch": 3.428367783321454,
3699
+ "grad_norm": 4.875,
3700
+ "learning_rate": 4.438516836417994e-07,
3701
+ "loss": 0.8649,
3702
+ "step": 4810
3703
+ },
3704
+ {
3705
+ "epoch": 3.4354953670705632,
3706
+ "grad_norm": 4.5625,
3707
+ "learning_rate": 3.986519777127884e-07,
3708
+ "loss": 0.7394,
3709
+ "step": 4820
3710
+ },
3711
+ {
3712
+ "epoch": 3.442622950819672,
3713
+ "grad_norm": 6.3125,
3714
+ "learning_rate": 3.558697715093207e-07,
3715
+ "loss": 0.6994,
3716
+ "step": 4830
3717
+ },
3718
+ {
3719
+ "epoch": 3.449750534568781,
3720
+ "grad_norm": 6.6875,
3721
+ "learning_rate": 3.1550715017575895e-07,
3722
+ "loss": 0.7571,
3723
+ "step": 4840
3724
+ },
3725
+ {
3726
+ "epoch": 3.4568781183178903,
3727
+ "grad_norm": 5.4375,
3728
+ "learning_rate": 2.775660809293368e-07,
3729
+ "loss": 0.6913,
3730
+ "step": 4850
3731
+ },
3732
+ {
3733
+ "epoch": 3.4640057020669994,
3734
+ "grad_norm": 5.90625,
3735
+ "learning_rate": 2.4204841296424085e-07,
3736
+ "loss": 0.7983,
3737
+ "step": 4860
3738
+ },
3739
+ {
3740
+ "epoch": 3.4711332858161086,
3741
+ "grad_norm": 5.46875,
3742
+ "learning_rate": 2.0895587736149414e-07,
3743
+ "loss": 0.7831,
3744
+ "step": 4870
3745
+ },
3746
+ {
3747
+ "epoch": 3.4782608695652173,
3748
+ "grad_norm": 6.21875,
3749
+ "learning_rate": 1.7829008700460115e-07,
3750
+ "loss": 0.7475,
3751
+ "step": 4880
3752
+ },
3753
+ {
3754
+ "epoch": 3.4853884533143265,
3755
+ "grad_norm": 4.9375,
3756
+ "learning_rate": 1.500525365009109e-07,
3757
+ "loss": 0.7402,
3758
+ "step": 4890
3759
+ },
3760
+ {
3761
+ "epoch": 3.4925160370634356,
3762
+ "grad_norm": 4.59375,
3763
+ "learning_rate": 1.2424460210881395e-07,
3764
+ "loss": 0.7151,
3765
+ "step": 4900
3766
+ },
3767
+ {
3768
+ "epoch": 3.4996436208125443,
3769
+ "grad_norm": 5.0,
3770
+ "learning_rate": 1.008675416706073e-07,
3771
+ "loss": 0.7341,
3772
+ "step": 4910
3773
+ },
3774
+ {
3775
+ "epoch": 3.5067712045616535,
3776
+ "grad_norm": 5.9375,
3777
+ "learning_rate": 7.99224945512489e-08,
3778
+ "loss": 0.717,
3779
+ "step": 4920
3780
+ },
3781
+ {
3782
+ "epoch": 3.5138987883107626,
3783
+ "grad_norm": 5.65625,
3784
+ "learning_rate": 6.141048158277429e-08,
3785
+ "loss": 0.7112,
3786
+ "step": 4930
3787
+ },
3788
+ {
3789
+ "epoch": 3.521026372059872,
3790
+ "grad_norm": 5.625,
3791
+ "learning_rate": 4.5332405014592014e-08,
3792
+ "loss": 0.747,
3793
+ "step": 4940
3794
+ },
3795
+ {
3796
+ "epoch": 3.528153955808981,
3797
+ "grad_norm": 5.65625,
3798
+ "learning_rate": 3.168904846945764e-08,
3799
+ "loss": 0.8131,
3800
+ "step": 4950
3801
+ },
3802
+ {
3803
+ "epoch": 3.5352815395580897,
3804
+ "grad_norm": 5.34375,
3805
+ "learning_rate": 2.0481076905332075e-08,
3806
+ "loss": 0.7257,
3807
+ "step": 4960
3808
+ },
3809
+ {
3810
+ "epoch": 3.542409123307199,
3811
+ "grad_norm": 6.34375,
3812
+ "learning_rate": 1.1709036582935318e-08,
3813
+ "loss": 0.7629,
3814
+ "step": 4970
3815
+ },
3816
+ {
3817
+ "epoch": 3.549536707056308,
3818
+ "grad_norm": 4.21875,
3819
+ "learning_rate": 5.373355039128835e-09,
3820
+ "loss": 0.681,
3821
+ "step": 4980
3822
+ },
3823
+ {
3824
+ "epoch": 3.556664290805417,
3825
+ "grad_norm": 6.71875,
3826
+ "learning_rate": 1.4743410661044455e-09,
3827
+ "loss": 0.8106,
3828
+ "step": 4990
3829
+ },
3830
+ {
3831
+ "epoch": 3.563791874554526,
3832
+ "grad_norm": 5.8125,
3833
+ "learning_rate": 1.2184696296380082e-11,
3834
+ "loss": 0.812,
3835
+ "step": 5000
3836
+ }
3837
+ ],
3838
+ "logging_steps": 10,
3839
+ "max_steps": 5000,
3840
+ "num_input_tokens_seen": 0,
3841
+ "num_train_epochs": 4,
3842
+ "save_steps": 1000,
3843
+ "stateful_callbacks": {
3844
+ "TrainerControl": {
3845
+ "args": {
3846
+ "should_epoch_stop": false,
3847
+ "should_evaluate": false,
3848
+ "should_log": false,
3849
+ "should_save": true,
3850
+ "should_training_stop": true
3851
+ },
3852
+ "attributes": {}
3853
+ }
3854
+ },
3855
+ "total_flos": 0.0,
3856
+ "train_batch_size": 16,
3857
+ "trial_name": null,
3858
+ "trial_params": null
3859
+ }