DylanJHJ commited on
Commit
d029c60
·
1 Parent(s): f64ab17

update the one trained with request

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +70 -0
  2. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/config.json +45 -0
  3. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/model.safetensors +3 -0
  4. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt +3 -0
  5. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_0.pth +3 -0
  6. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_1.pth +3 -0
  7. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_2.pth +3 -0
  8. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_3.pth +3 -0
  9. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/scheduler.pt +3 -0
  10. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/trainer_state.json +799 -0
  11. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/training_args.bin +3 -0
  12. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/config.json +45 -0
  13. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/model.safetensors +3 -0
  14. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt +3 -0
  15. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_0.pth +3 -0
  16. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_1.pth +3 -0
  17. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_2.pth +3 -0
  18. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_3.pth +3 -0
  19. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/scheduler.pt +3 -0
  20. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/trainer_state.json +1564 -0
  21. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/training_args.bin +3 -0
  22. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/config.json +45 -0
  23. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/model.safetensors +3 -0
  24. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt +3 -0
  25. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_0.pth +3 -0
  26. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_1.pth +3 -0
  27. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_2.pth +3 -0
  28. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_3.pth +3 -0
  29. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/scheduler.pt +3 -0
  30. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/trainer_state.json +2329 -0
  31. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/training_args.bin +3 -0
  32. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/config.json +45 -0
  33. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/model.safetensors +3 -0
  34. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt +3 -0
  35. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_0.pth +3 -0
  36. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_1.pth +3 -0
  37. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_2.pth +3 -0
  38. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_3.pth +3 -0
  39. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/scheduler.pt +3 -0
  40. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/trainer_state.json +3094 -0
  41. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/training_args.bin +3 -0
  42. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-5000/moved +0 -0
  43. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/config.json +45 -0
  44. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/model.safetensors +3 -0
  45. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/optimizer.pt +3 -0
  46. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_0.pth +3 -0
  47. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_1.pth +3 -0
  48. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_2.pth +3 -0
  49. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_3.pth +3 -0
  50. modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/runs/Jan01_22-56-07_nid005051/events.out.tfevents.1767301064.nid005051.19581.0 +3 -0
.gitattributes CHANGED
@@ -350,3 +350,73 @@ modernbert-crux-researchy-flatten.b64_n512.1e-4/optimizer.pt filter=lfs diff=lfs
350
  .git/lfs/objects/3d/4a/3d4a3a4394bc050534ccc08756a61ea984b16584719563b3ae55a8471b6fa937 filter=lfs diff=lfs merge=lfs -text
351
  .git/lfs/objects/fd/82/fd82dce9a1914d1427efa2b58c72a13ee36999640c614adaa63b2f789fe49c72 filter=lfs diff=lfs merge=lfs -text
352
  .git/lfs/objects/92/0f/920f334aa8a2a64cf006d7a05ba27b3b53bb416a0f30df954d0ba598a097fa3b filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
350
  .git/lfs/objects/3d/4a/3d4a3a4394bc050534ccc08756a61ea984b16584719563b3ae55a8471b6fa937 filter=lfs diff=lfs merge=lfs -text
351
  .git/lfs/objects/fd/82/fd82dce9a1914d1427efa2b58c72a13ee36999640c614adaa63b2f789fe49c72 filter=lfs diff=lfs merge=lfs -text
352
  .git/lfs/objects/92/0f/920f334aa8a2a64cf006d7a05ba27b3b53bb416a0f30df954d0ba598a097fa3b filter=lfs diff=lfs merge=lfs -text
353
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
354
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
355
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
356
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
357
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
358
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
359
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
360
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
361
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
362
+ modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
363
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
364
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
365
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
366
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
367
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
368
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
369
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
370
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
371
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
372
+ modernbert-crux-researchy-pos_high.neg_zero.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
373
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
374
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
375
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
376
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
377
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
378
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
379
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
380
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
381
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
382
+ modernbert-crux-researchy-pos_half.neg_zero.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
383
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
384
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
385
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
386
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
387
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
388
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
389
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
390
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
391
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
392
+ modernbert-crux-researchy-pos_zero.neg_high.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
393
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
394
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
395
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
396
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
397
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
398
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
399
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
400
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
401
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
402
+ modernbert-crux-researchy-pos_high.neg_low.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
403
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
404
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
405
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
406
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
407
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
408
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
409
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
410
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
411
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
412
+ modernbert-crux-researchy-pos_high.neg_quarter.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
413
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
414
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
415
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
416
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
417
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/model.safetensors filter=lfs diff=lfs merge=lfs -text
418
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/optimizer.pt filter=lfs diff=lfs merge=lfs -text
419
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-1000/model.safetensors filter=lfs diff=lfs merge=lfs -text
420
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
421
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-4000/model.safetensors filter=lfs diff=lfs merge=lfs -text
422
+ modernbert-crux-researchy-pos_low.neg_zero.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca7454cf27ca9f5516aa6fff961b75093665ac6e1dc9b38172ef9865db0763e2
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47e732067fbfe03b9b9d1db6e092429169ba6dee3aa99b55786815e28b2ed522
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3044a568f2c48f55afab6e63cf0c046d3586784e77e39b1af825abecfb2cdba8
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86accf27064cdd503053e90476a6bd10de333d4ff0594535ad55ea13a473c91d
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18ca8d714ef40be035404c1957b5a4dee84e1f43980408393f8aa710552ee6f6
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cfdebe99e40accc9c9d8f09c63136a14abda997d9b501969ec8e16e9d183179
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:053b181ac2db5ceae48310494587c492212456802136bc3852c62ccc7b9ff38c
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,799 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.7132667617689016,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007132667617689016,
79
+ "grad_norm": 18.375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.6901,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014265335235378032,
86
+ "grad_norm": 10.75,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.7675,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021398002853067047,
93
+ "grad_norm": 21.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.527,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.028530670470756064,
100
+ "grad_norm": 19.0,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.7708,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03566333808844508,
107
+ "grad_norm": 17.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.4093,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042796005706134094,
114
+ "grad_norm": 15.625,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.579,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04992867332382311,
121
+ "grad_norm": 14.0,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.4845,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05706134094151213,
128
+ "grad_norm": 13.6875,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.598,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06419400855920114,
135
+ "grad_norm": 15.5625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3226,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07132667617689016,
142
+ "grad_norm": 13.8125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.4969,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07845934379457917,
149
+ "grad_norm": 13.125,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3867,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08559201141226819,
156
+ "grad_norm": 14.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.1904,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09272467902995721,
163
+ "grad_norm": 9.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1668,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09985734664764621,
170
+ "grad_norm": 8.1875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.124,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10699001426533523,
177
+ "grad_norm": 9.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.1571,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11412268188302425,
184
+ "grad_norm": 12.25,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2145,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12125534950071326,
191
+ "grad_norm": 10.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 1.155,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12838801711840228,
198
+ "grad_norm": 8.5,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1513,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.1355206847360913,
205
+ "grad_norm": 9.5,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.9803,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14265335235378032,
212
+ "grad_norm": 8.0625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.0302,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.14978601997146934,
219
+ "grad_norm": 9.25,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.9878,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15691868758915833,
226
+ "grad_norm": 10.25,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9816,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16405135520684735,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0415,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17118402282453637,
240
+ "grad_norm": 10.1875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.1126,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1783166904422254,
247
+ "grad_norm": 7.78125,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0127,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18544935805991442,
254
+ "grad_norm": 5.71875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9577,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.19258202567760344,
261
+ "grad_norm": 6.21875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.9518,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19971469329529243,
268
+ "grad_norm": 6.28125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7952,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.20684736091298145,
275
+ "grad_norm": 6.875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9201,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21398002853067047,
282
+ "grad_norm": 6.09375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.951,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2211126961483595,
289
+ "grad_norm": 7.65625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.9199,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.2282453637660485,
296
+ "grad_norm": 7.3125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.8773,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23537803138373753,
303
+ "grad_norm": 4.875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.9386,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24251069900142652,
310
+ "grad_norm": 5.1875,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9364,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.24964336661911554,
317
+ "grad_norm": 7.21875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9053,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25677603423680456,
324
+ "grad_norm": 5.8125,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.8493,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.26390870185449355,
331
+ "grad_norm": 7.1875,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8218,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.2710413694721826,
338
+ "grad_norm": 5.625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.8975,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.2781740370898716,
345
+ "grad_norm": 5.1875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8891,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.28530670470756064,
352
+ "grad_norm": 5.78125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.7831,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29243937232524964,
359
+ "grad_norm": 5.125,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9017,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2995720399429387,
366
+ "grad_norm": 4.96875,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.7839,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.3067047075606277,
373
+ "grad_norm": 5.5625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.886,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.31383737517831667,
380
+ "grad_norm": 6.1875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.863,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3209700427960057,
387
+ "grad_norm": 4.625,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.884,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.3281027104136947,
394
+ "grad_norm": 5.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9209,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33523537803138376,
401
+ "grad_norm": 6.09375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8775,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.34236804564907275,
408
+ "grad_norm": 5.8125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.7956,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34950071326676174,
415
+ "grad_norm": 5.71875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8632,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3566333808844508,
422
+ "grad_norm": 4.4375,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.8308,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.3637660485021398,
429
+ "grad_norm": 5.25,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.884,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.37089871611982883,
436
+ "grad_norm": 5.28125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.8663,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.3780313837375178,
443
+ "grad_norm": 5.25,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.8722,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.38516405135520687,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8391,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.39229671897289586,
457
+ "grad_norm": 5.5625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.7823,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39942938659058486,
464
+ "grad_norm": 4.875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8606,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4065620542082739,
471
+ "grad_norm": 5.625,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8719,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4136947218259629,
478
+ "grad_norm": 4.71875,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.8619,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42082738944365194,
485
+ "grad_norm": 5.0,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.8439,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.42796005706134094,
492
+ "grad_norm": 5.0625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.886,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43509272467902993,
499
+ "grad_norm": 5.125,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.8346,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.442225392296719,
506
+ "grad_norm": 5.40625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.8145,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44935805991440797,
513
+ "grad_norm": 5.0625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.8739,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.456490727532097,
520
+ "grad_norm": 4.75,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.7826,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.463623395149786,
527
+ "grad_norm": 5.6875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8541,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47075606276747506,
534
+ "grad_norm": 4.3125,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.8208,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.47788873038516405,
541
+ "grad_norm": 5.375,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7761,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48502139800285304,
548
+ "grad_norm": 5.15625,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8346,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4921540656205421,
555
+ "grad_norm": 5.03125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.9192,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4992867332382311,
562
+ "grad_norm": 4.96875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.8037,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5064194008559201,
569
+ "grad_norm": 4.53125,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.7669,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5135520684736091,
576
+ "grad_norm": 6.875,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.7915,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5206847360912982,
583
+ "grad_norm": 5.0625,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.7863,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5278174037089871,
590
+ "grad_norm": 5.21875,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.7789,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5349500713266762,
597
+ "grad_norm": 4.0,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.7759,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5420827389443652,
604
+ "grad_norm": 5.34375,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.8504,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5492154065620543,
611
+ "grad_norm": 4.59375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.809,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5563480741797432,
618
+ "grad_norm": 5.53125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.7721,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5634807417974322,
625
+ "grad_norm": 4.8125,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.7319,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5706134094151213,
632
+ "grad_norm": 4.84375,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7726,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5777460770328102,
639
+ "grad_norm": 4.625,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8522,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5848787446504993,
646
+ "grad_norm": 4.34375,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.7761,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5920114122681883,
653
+ "grad_norm": 5.0625,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8702,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5991440798858774,
660
+ "grad_norm": 5.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8636,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6062767475035663,
667
+ "grad_norm": 3.96875,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6974,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6134094151212554,
674
+ "grad_norm": 5.03125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.8017,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6205420827389444,
681
+ "grad_norm": 5.40625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7799,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6276747503566333,
688
+ "grad_norm": 5.125,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.7808,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6348074179743224,
695
+ "grad_norm": 5.4375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8337,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6419400855920114,
702
+ "grad_norm": 4.90625,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.8043,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6490727532097005,
709
+ "grad_norm": 4.59375,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.8328,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6562054208273894,
716
+ "grad_norm": 4.59375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7745,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6633380884450785,
723
+ "grad_norm": 5.5625,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.745,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6704707560627675,
730
+ "grad_norm": 4.96875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.82,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.6776034236804565,
737
+ "grad_norm": 4.78125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8025,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.6847360912981455,
744
+ "grad_norm": 5.90625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.84,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6918687589158345,
751
+ "grad_norm": 4.21875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.8552,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6990014265335235,
758
+ "grad_norm": 4.15625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.8082,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7061340941512125,
765
+ "grad_norm": 4.71875,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8265,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7132667617689016,
772
+ "grad_norm": 4.875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8369,
775
+ "step": 1000
776
+ }
777
+ ],
778
+ "logging_steps": 10,
779
+ "max_steps": 5000,
780
+ "num_input_tokens_seen": 0,
781
+ "num_train_epochs": 4,
782
+ "save_steps": 1000,
783
+ "stateful_callbacks": {
784
+ "TrainerControl": {
785
+ "args": {
786
+ "should_epoch_stop": false,
787
+ "should_evaluate": false,
788
+ "should_log": false,
789
+ "should_save": true,
790
+ "should_training_stop": false
791
+ },
792
+ "attributes": {}
793
+ }
794
+ },
795
+ "total_flos": 0.0,
796
+ "train_batch_size": 16,
797
+ "trial_name": null,
798
+ "trial_params": null
799
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af7cf7c3c741add5e8996639d61d87fe74bb40ad0faa1b29901130f4b3a60fc3
3
+ size 6353
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:096d65b355d6e0613adaabb2f411c1c64396ffbf43bdd18442af55813a089b23
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:944f7393a46576d6228b50f137c797dcfc63520740b5d5942abe6fffc0c37b37
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18bb6b6e7da6250d22c25b95c64287be4aea598d9d97ef67b0fd69b9be869ed7
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13db4dd10b22b5464f5ef0b539152b46390eae6a827b355108cb54b1af2eb830
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06e293f13fb6e873e8081636b6b07115410b397c55f1ed08b8772b059c8eb74e
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c595c077664f94b7f91088561951be20e39483e01865774a83bc46f3971a3152
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff65ca8b6e225d1d82cc63887ee464e8c896624115a4dc693eae1fff20815699
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/trainer_state.json ADDED
@@ -0,0 +1,1564 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.4265335235378032,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 2000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007132667617689016,
79
+ "grad_norm": 18.375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.6901,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014265335235378032,
86
+ "grad_norm": 10.75,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.7675,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021398002853067047,
93
+ "grad_norm": 21.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.527,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.028530670470756064,
100
+ "grad_norm": 19.0,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.7708,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03566333808844508,
107
+ "grad_norm": 17.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.4093,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042796005706134094,
114
+ "grad_norm": 15.625,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.579,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04992867332382311,
121
+ "grad_norm": 14.0,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.4845,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05706134094151213,
128
+ "grad_norm": 13.6875,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.598,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06419400855920114,
135
+ "grad_norm": 15.5625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3226,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07132667617689016,
142
+ "grad_norm": 13.8125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.4969,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07845934379457917,
149
+ "grad_norm": 13.125,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3867,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08559201141226819,
156
+ "grad_norm": 14.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.1904,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09272467902995721,
163
+ "grad_norm": 9.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1668,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09985734664764621,
170
+ "grad_norm": 8.1875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.124,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10699001426533523,
177
+ "grad_norm": 9.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.1571,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11412268188302425,
184
+ "grad_norm": 12.25,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2145,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12125534950071326,
191
+ "grad_norm": 10.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 1.155,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12838801711840228,
198
+ "grad_norm": 8.5,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1513,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.1355206847360913,
205
+ "grad_norm": 9.5,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.9803,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14265335235378032,
212
+ "grad_norm": 8.0625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.0302,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.14978601997146934,
219
+ "grad_norm": 9.25,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.9878,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15691868758915833,
226
+ "grad_norm": 10.25,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9816,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16405135520684735,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0415,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17118402282453637,
240
+ "grad_norm": 10.1875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.1126,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1783166904422254,
247
+ "grad_norm": 7.78125,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0127,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18544935805991442,
254
+ "grad_norm": 5.71875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9577,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.19258202567760344,
261
+ "grad_norm": 6.21875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.9518,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19971469329529243,
268
+ "grad_norm": 6.28125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7952,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.20684736091298145,
275
+ "grad_norm": 6.875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9201,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21398002853067047,
282
+ "grad_norm": 6.09375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.951,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2211126961483595,
289
+ "grad_norm": 7.65625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.9199,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.2282453637660485,
296
+ "grad_norm": 7.3125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.8773,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23537803138373753,
303
+ "grad_norm": 4.875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.9386,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24251069900142652,
310
+ "grad_norm": 5.1875,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9364,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.24964336661911554,
317
+ "grad_norm": 7.21875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9053,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25677603423680456,
324
+ "grad_norm": 5.8125,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.8493,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.26390870185449355,
331
+ "grad_norm": 7.1875,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8218,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.2710413694721826,
338
+ "grad_norm": 5.625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.8975,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.2781740370898716,
345
+ "grad_norm": 5.1875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8891,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.28530670470756064,
352
+ "grad_norm": 5.78125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.7831,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29243937232524964,
359
+ "grad_norm": 5.125,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9017,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2995720399429387,
366
+ "grad_norm": 4.96875,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.7839,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.3067047075606277,
373
+ "grad_norm": 5.5625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.886,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.31383737517831667,
380
+ "grad_norm": 6.1875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.863,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3209700427960057,
387
+ "grad_norm": 4.625,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.884,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.3281027104136947,
394
+ "grad_norm": 5.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9209,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33523537803138376,
401
+ "grad_norm": 6.09375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8775,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.34236804564907275,
408
+ "grad_norm": 5.8125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.7956,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34950071326676174,
415
+ "grad_norm": 5.71875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8632,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3566333808844508,
422
+ "grad_norm": 4.4375,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.8308,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.3637660485021398,
429
+ "grad_norm": 5.25,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.884,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.37089871611982883,
436
+ "grad_norm": 5.28125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.8663,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.3780313837375178,
443
+ "grad_norm": 5.25,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.8722,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.38516405135520687,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8391,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.39229671897289586,
457
+ "grad_norm": 5.5625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.7823,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39942938659058486,
464
+ "grad_norm": 4.875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8606,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4065620542082739,
471
+ "grad_norm": 5.625,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8719,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4136947218259629,
478
+ "grad_norm": 4.71875,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.8619,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42082738944365194,
485
+ "grad_norm": 5.0,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.8439,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.42796005706134094,
492
+ "grad_norm": 5.0625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.886,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43509272467902993,
499
+ "grad_norm": 5.125,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.8346,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.442225392296719,
506
+ "grad_norm": 5.40625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.8145,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44935805991440797,
513
+ "grad_norm": 5.0625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.8739,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.456490727532097,
520
+ "grad_norm": 4.75,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.7826,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.463623395149786,
527
+ "grad_norm": 5.6875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8541,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47075606276747506,
534
+ "grad_norm": 4.3125,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.8208,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.47788873038516405,
541
+ "grad_norm": 5.375,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7761,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48502139800285304,
548
+ "grad_norm": 5.15625,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8346,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4921540656205421,
555
+ "grad_norm": 5.03125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.9192,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4992867332382311,
562
+ "grad_norm": 4.96875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.8037,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5064194008559201,
569
+ "grad_norm": 4.53125,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.7669,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5135520684736091,
576
+ "grad_norm": 6.875,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.7915,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5206847360912982,
583
+ "grad_norm": 5.0625,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.7863,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5278174037089871,
590
+ "grad_norm": 5.21875,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.7789,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5349500713266762,
597
+ "grad_norm": 4.0,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.7759,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5420827389443652,
604
+ "grad_norm": 5.34375,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.8504,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5492154065620543,
611
+ "grad_norm": 4.59375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.809,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5563480741797432,
618
+ "grad_norm": 5.53125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.7721,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5634807417974322,
625
+ "grad_norm": 4.8125,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.7319,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5706134094151213,
632
+ "grad_norm": 4.84375,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7726,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5777460770328102,
639
+ "grad_norm": 4.625,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8522,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5848787446504993,
646
+ "grad_norm": 4.34375,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.7761,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5920114122681883,
653
+ "grad_norm": 5.0625,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8702,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5991440798858774,
660
+ "grad_norm": 5.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8636,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6062767475035663,
667
+ "grad_norm": 3.96875,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6974,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6134094151212554,
674
+ "grad_norm": 5.03125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.8017,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6205420827389444,
681
+ "grad_norm": 5.40625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7799,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6276747503566333,
688
+ "grad_norm": 5.125,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.7808,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6348074179743224,
695
+ "grad_norm": 5.4375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8337,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6419400855920114,
702
+ "grad_norm": 4.90625,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.8043,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6490727532097005,
709
+ "grad_norm": 4.59375,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.8328,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6562054208273894,
716
+ "grad_norm": 4.59375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7745,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6633380884450785,
723
+ "grad_norm": 5.5625,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.745,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6704707560627675,
730
+ "grad_norm": 4.96875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.82,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.6776034236804565,
737
+ "grad_norm": 4.78125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8025,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.6847360912981455,
744
+ "grad_norm": 5.90625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.84,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6918687589158345,
751
+ "grad_norm": 4.21875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.8552,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6990014265335235,
758
+ "grad_norm": 4.15625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.8082,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7061340941512125,
765
+ "grad_norm": 4.71875,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8265,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7132667617689016,
772
+ "grad_norm": 4.875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8369,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7132667617689016,
779
+ "eval/nano_beir.arguana": 0.47952715537701884,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7132667617689016,
784
+ "eval/nano_beir.climate_fever": 0.22702207934204544,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7132667617689016,
789
+ "eval/nano_beir.dbpedia_entity": 0.3287396364958694,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7132667617689016,
794
+ "eval/nano_beir.fever": 0.5532628324042129,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7132667617689016,
799
+ "eval/nano_beir.fiqa": 0.4160855038689836,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7132667617689016,
804
+ "eval/nano_beir.hotpotqa": 0.6599716759215029,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7132667617689016,
809
+ "eval/nano_beir.nfcorpus": 0.3449425020579428,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7132667617689016,
814
+ "eval/nano_beir.nq": 0.39316532973343116,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7132667617689016,
819
+ "eval/nano_beir.quora": 0.8295275476072758,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7132667617689016,
824
+ "eval/nano_beir.scidocs": 0.2526832556333601,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7132667617689016,
829
+ "eval/nano_beir.scifact": 0.7219633063518482,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7132667617689016,
834
+ "eval/nano_beir.webis_touche2020": 0.2787587910246796,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7132667617689016,
839
+ "eval/avg": 0.4571374679848476,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7203994293865906,
844
+ "grad_norm": 5.15625,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8872,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7275320970042796,
851
+ "grad_norm": 5.65625,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8334,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7346647646219686,
858
+ "grad_norm": 5.1875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8014,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7417974322396577,
865
+ "grad_norm": 4.9375,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.848,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7489300998573466,
872
+ "grad_norm": 4.40625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.7203,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7560627674750356,
879
+ "grad_norm": 5.0625,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8405,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7631954350927247,
886
+ "grad_norm": 4.5625,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.8418,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7703281027104137,
893
+ "grad_norm": 5.75,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.8853,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7774607703281027,
900
+ "grad_norm": 4.09375,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.7733,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7845934379457917,
907
+ "grad_norm": 4.40625,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.7698,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7917261055634808,
914
+ "grad_norm": 4.84375,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.7759,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7988587731811697,
921
+ "grad_norm": 5.90625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8025,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8059914407988588,
928
+ "grad_norm": 5.53125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.8364,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8131241084165478,
935
+ "grad_norm": 5.03125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.7523,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.8202567760342369,
942
+ "grad_norm": 5.46875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7333,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.8273894436519258,
949
+ "grad_norm": 4.875,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.7325,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8345221112696148,
956
+ "grad_norm": 5.0,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7996,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8416547788873039,
963
+ "grad_norm": 4.6875,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7946,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8487874465049928,
970
+ "grad_norm": 5.21875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.9133,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8559201141226819,
977
+ "grad_norm": 4.78125,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.7218,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8630527817403709,
984
+ "grad_norm": 4.28125,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8453,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8701854493580599,
991
+ "grad_norm": 5.1875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.7261,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8773181169757489,
998
+ "grad_norm": 3.828125,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.6459,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.884450784593438,
1005
+ "grad_norm": 5.3125,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.7202,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.891583452211127,
1012
+ "grad_norm": 5.90625,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.7874,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8987161198288159,
1019
+ "grad_norm": 4.09375,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8524,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.905848787446505,
1026
+ "grad_norm": 5.75,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.7854,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.912981455064194,
1033
+ "grad_norm": 5.3125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8386,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.920114122681883,
1040
+ "grad_norm": 4.71875,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.8279,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.927246790299572,
1047
+ "grad_norm": 4.71875,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.8345,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9343794579172611,
1054
+ "grad_norm": 5.15625,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7596,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9415121255349501,
1061
+ "grad_norm": 4.4375,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.6704,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.948644793152639,
1068
+ "grad_norm": 5.6875,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7792,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.9557774607703281,
1075
+ "grad_norm": 4.78125,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.813,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.9629101283880172,
1082
+ "grad_norm": 4.4375,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.834,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9700427960057061,
1089
+ "grad_norm": 4.1875,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.8299,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9771754636233951,
1096
+ "grad_norm": 4.3125,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.7407,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9843081312410842,
1103
+ "grad_norm": 5.0,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.8821,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9914407988587732,
1110
+ "grad_norm": 4.71875,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8595,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9985734664764622,
1117
+ "grad_norm": 4.5,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.869,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.005706134094151,
1124
+ "grad_norm": 4.28125,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.8397,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0128388017118402,
1131
+ "grad_norm": 5.34375,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8047,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0199714693295292,
1138
+ "grad_norm": 4.28125,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.72,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0271041369472182,
1145
+ "grad_norm": 5.59375,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.7724,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0342368045649073,
1152
+ "grad_norm": 4.3125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.8014,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0413694721825963,
1159
+ "grad_norm": 6.03125,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.8532,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0485021398002854,
1166
+ "grad_norm": 3.75,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.7285,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0556348074179742,
1173
+ "grad_norm": 4.5625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.8574,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0627674750356633,
1180
+ "grad_norm": 6.21875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.76,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0699001426533523,
1187
+ "grad_norm": 4.75,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.8441,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0770328102710414,
1194
+ "grad_norm": 4.5625,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.764,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0841654778887304,
1201
+ "grad_norm": 4.59375,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.7432,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.0912981455064195,
1208
+ "grad_norm": 4.1875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.7894,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0984308131241085,
1215
+ "grad_norm": 4.5,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.761,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.1055634807417973,
1222
+ "grad_norm": 6.03125,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.9432,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1126961483594864,
1229
+ "grad_norm": 4.1875,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7863,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1198288159771754,
1236
+ "grad_norm": 4.21875,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.7732,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1269614835948645,
1243
+ "grad_norm": 4.65625,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.7651,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1340941512125535,
1250
+ "grad_norm": 4.34375,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8021,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1412268188302426,
1257
+ "grad_norm": 5.0625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8561,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1483594864479316,
1264
+ "grad_norm": 5.03125,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7551,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1554921540656204,
1271
+ "grad_norm": 5.375,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.8045,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1626248216833095,
1278
+ "grad_norm": 4.21875,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.7145,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1697574893009985,
1285
+ "grad_norm": 4.125,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8007,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1768901569186876,
1292
+ "grad_norm": 5.03125,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8009,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1840228245363766,
1299
+ "grad_norm": 4.84375,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.833,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1911554921540657,
1306
+ "grad_norm": 5.21875,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8149,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1982881597717547,
1313
+ "grad_norm": 4.1875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.808,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2054208273894436,
1320
+ "grad_norm": 4.625,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7549,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2125534950071326,
1327
+ "grad_norm": 5.71875,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7755,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.2196861626248217,
1334
+ "grad_norm": 4.5625,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8038,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.2268188302425107,
1341
+ "grad_norm": 4.09375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.7207,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.2339514978601998,
1348
+ "grad_norm": 4.90625,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7994,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.2410841654778888,
1355
+ "grad_norm": 4.3125,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7989,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2482168330955776,
1362
+ "grad_norm": 4.34375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.6516,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2553495007132667,
1369
+ "grad_norm": 4.59375,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7919,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2624821683309557,
1376
+ "grad_norm": 5.53125,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.9372,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2696148359486448,
1383
+ "grad_norm": 4.3125,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7683,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2767475035663338,
1390
+ "grad_norm": 4.3125,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.737,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2838801711840229,
1397
+ "grad_norm": 5.78125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.7507,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.291012838801712,
1404
+ "grad_norm": 4.875,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7016,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.298145506419401,
1411
+ "grad_norm": 4.5,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7151,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3052781740370898,
1418
+ "grad_norm": 4.5625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8559,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3124108416547788,
1425
+ "grad_norm": 4.15625,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7118,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3195435092724679,
1432
+ "grad_norm": 3.78125,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.7855,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.326676176890157,
1439
+ "grad_norm": 4.53125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.7438,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.333808844507846,
1446
+ "grad_norm": 4.65625,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.7018,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.340941512125535,
1453
+ "grad_norm": 6.9375,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.7083,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3480741797432239,
1460
+ "grad_norm": 4.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.8265,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.355206847360913,
1467
+ "grad_norm": 5.5625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.8735,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.362339514978602,
1474
+ "grad_norm": 5.71875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.7256,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.369472182596291,
1481
+ "grad_norm": 4.53125,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.7472,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.37660485021398,
1488
+ "grad_norm": 5.875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7601,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.383737517831669,
1495
+ "grad_norm": 5.75,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8857,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.3908701854493581,
1502
+ "grad_norm": 6.40625,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7296,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3980028530670472,
1509
+ "grad_norm": 4.40625,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7331,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.405135520684736,
1516
+ "grad_norm": 5.09375,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.8462,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.412268188302425,
1523
+ "grad_norm": 4.15625,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.7307,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4194008559201141,
1530
+ "grad_norm": 4.90625,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7711,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4265335235378032,
1537
+ "grad_norm": 5.78125,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.7485,
1540
+ "step": 2000
1541
+ }
1542
+ ],
1543
+ "logging_steps": 10,
1544
+ "max_steps": 5000,
1545
+ "num_input_tokens_seen": 0,
1546
+ "num_train_epochs": 4,
1547
+ "save_steps": 1000,
1548
+ "stateful_callbacks": {
1549
+ "TrainerControl": {
1550
+ "args": {
1551
+ "should_epoch_stop": false,
1552
+ "should_evaluate": false,
1553
+ "should_log": false,
1554
+ "should_save": true,
1555
+ "should_training_stop": false
1556
+ },
1557
+ "attributes": {}
1558
+ }
1559
+ },
1560
+ "total_flos": 0.0,
1561
+ "train_batch_size": 16,
1562
+ "trial_name": null,
1563
+ "trial_params": null
1564
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-2000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af7cf7c3c741add5e8996639d61d87fe74bb40ad0faa1b29901130f4b3a60fc3
3
+ size 6353
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73bd96a7a178cbb1b6042cb786b17db797070f08fb5cdd2b8c9292b9919e8ea6
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65d16bc47175a6c00670f9468b6365667a23c8eaf3d7b5528554ef44bf7728ee
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0e770c61e3acef7bd144efee65a450ed487b244ed293a8dd801394ffcae775
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fb125336725f7741cb4daa1e3d06e225bbacfde8d41c4dcabb6762c222e62c6
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153c865f77c7129ba565bded50f334683d51c80f20e3cfec39e62f8737b86f0d
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d017ce00fcebac7edc058ddd138f194eb0340f2d8ad0879bdab08f922ed0846e
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a8f419d963b829163d14296852a07ad341e5e63e49d44de7c0a10c9369d217
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/trainer_state.json ADDED
@@ -0,0 +1,2329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.1398002853067046,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 3000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007132667617689016,
79
+ "grad_norm": 18.375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.6901,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014265335235378032,
86
+ "grad_norm": 10.75,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.7675,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021398002853067047,
93
+ "grad_norm": 21.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.527,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.028530670470756064,
100
+ "grad_norm": 19.0,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.7708,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03566333808844508,
107
+ "grad_norm": 17.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.4093,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042796005706134094,
114
+ "grad_norm": 15.625,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.579,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04992867332382311,
121
+ "grad_norm": 14.0,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.4845,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05706134094151213,
128
+ "grad_norm": 13.6875,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.598,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06419400855920114,
135
+ "grad_norm": 15.5625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3226,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07132667617689016,
142
+ "grad_norm": 13.8125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.4969,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07845934379457917,
149
+ "grad_norm": 13.125,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3867,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08559201141226819,
156
+ "grad_norm": 14.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.1904,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09272467902995721,
163
+ "grad_norm": 9.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1668,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09985734664764621,
170
+ "grad_norm": 8.1875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.124,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10699001426533523,
177
+ "grad_norm": 9.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.1571,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11412268188302425,
184
+ "grad_norm": 12.25,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2145,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12125534950071326,
191
+ "grad_norm": 10.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 1.155,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12838801711840228,
198
+ "grad_norm": 8.5,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1513,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.1355206847360913,
205
+ "grad_norm": 9.5,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.9803,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14265335235378032,
212
+ "grad_norm": 8.0625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.0302,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.14978601997146934,
219
+ "grad_norm": 9.25,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.9878,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15691868758915833,
226
+ "grad_norm": 10.25,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9816,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16405135520684735,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0415,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17118402282453637,
240
+ "grad_norm": 10.1875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.1126,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1783166904422254,
247
+ "grad_norm": 7.78125,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0127,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18544935805991442,
254
+ "grad_norm": 5.71875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9577,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.19258202567760344,
261
+ "grad_norm": 6.21875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.9518,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19971469329529243,
268
+ "grad_norm": 6.28125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7952,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.20684736091298145,
275
+ "grad_norm": 6.875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9201,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21398002853067047,
282
+ "grad_norm": 6.09375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.951,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2211126961483595,
289
+ "grad_norm": 7.65625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.9199,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.2282453637660485,
296
+ "grad_norm": 7.3125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.8773,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23537803138373753,
303
+ "grad_norm": 4.875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.9386,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24251069900142652,
310
+ "grad_norm": 5.1875,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9364,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.24964336661911554,
317
+ "grad_norm": 7.21875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9053,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25677603423680456,
324
+ "grad_norm": 5.8125,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.8493,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.26390870185449355,
331
+ "grad_norm": 7.1875,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8218,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.2710413694721826,
338
+ "grad_norm": 5.625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.8975,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.2781740370898716,
345
+ "grad_norm": 5.1875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8891,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.28530670470756064,
352
+ "grad_norm": 5.78125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.7831,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29243937232524964,
359
+ "grad_norm": 5.125,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9017,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2995720399429387,
366
+ "grad_norm": 4.96875,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.7839,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.3067047075606277,
373
+ "grad_norm": 5.5625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.886,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.31383737517831667,
380
+ "grad_norm": 6.1875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.863,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3209700427960057,
387
+ "grad_norm": 4.625,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.884,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.3281027104136947,
394
+ "grad_norm": 5.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9209,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33523537803138376,
401
+ "grad_norm": 6.09375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8775,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.34236804564907275,
408
+ "grad_norm": 5.8125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.7956,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34950071326676174,
415
+ "grad_norm": 5.71875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8632,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3566333808844508,
422
+ "grad_norm": 4.4375,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.8308,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.3637660485021398,
429
+ "grad_norm": 5.25,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.884,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.37089871611982883,
436
+ "grad_norm": 5.28125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.8663,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.3780313837375178,
443
+ "grad_norm": 5.25,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.8722,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.38516405135520687,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8391,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.39229671897289586,
457
+ "grad_norm": 5.5625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.7823,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39942938659058486,
464
+ "grad_norm": 4.875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8606,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4065620542082739,
471
+ "grad_norm": 5.625,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8719,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4136947218259629,
478
+ "grad_norm": 4.71875,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.8619,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42082738944365194,
485
+ "grad_norm": 5.0,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.8439,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.42796005706134094,
492
+ "grad_norm": 5.0625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.886,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43509272467902993,
499
+ "grad_norm": 5.125,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.8346,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.442225392296719,
506
+ "grad_norm": 5.40625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.8145,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44935805991440797,
513
+ "grad_norm": 5.0625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.8739,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.456490727532097,
520
+ "grad_norm": 4.75,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.7826,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.463623395149786,
527
+ "grad_norm": 5.6875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8541,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47075606276747506,
534
+ "grad_norm": 4.3125,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.8208,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.47788873038516405,
541
+ "grad_norm": 5.375,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7761,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48502139800285304,
548
+ "grad_norm": 5.15625,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8346,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4921540656205421,
555
+ "grad_norm": 5.03125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.9192,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4992867332382311,
562
+ "grad_norm": 4.96875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.8037,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5064194008559201,
569
+ "grad_norm": 4.53125,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.7669,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5135520684736091,
576
+ "grad_norm": 6.875,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.7915,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5206847360912982,
583
+ "grad_norm": 5.0625,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.7863,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5278174037089871,
590
+ "grad_norm": 5.21875,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.7789,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5349500713266762,
597
+ "grad_norm": 4.0,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.7759,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5420827389443652,
604
+ "grad_norm": 5.34375,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.8504,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5492154065620543,
611
+ "grad_norm": 4.59375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.809,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5563480741797432,
618
+ "grad_norm": 5.53125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.7721,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5634807417974322,
625
+ "grad_norm": 4.8125,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.7319,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5706134094151213,
632
+ "grad_norm": 4.84375,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7726,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5777460770328102,
639
+ "grad_norm": 4.625,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8522,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5848787446504993,
646
+ "grad_norm": 4.34375,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.7761,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5920114122681883,
653
+ "grad_norm": 5.0625,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8702,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5991440798858774,
660
+ "grad_norm": 5.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8636,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6062767475035663,
667
+ "grad_norm": 3.96875,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6974,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6134094151212554,
674
+ "grad_norm": 5.03125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.8017,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6205420827389444,
681
+ "grad_norm": 5.40625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7799,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6276747503566333,
688
+ "grad_norm": 5.125,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.7808,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6348074179743224,
695
+ "grad_norm": 5.4375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8337,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6419400855920114,
702
+ "grad_norm": 4.90625,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.8043,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6490727532097005,
709
+ "grad_norm": 4.59375,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.8328,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6562054208273894,
716
+ "grad_norm": 4.59375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7745,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6633380884450785,
723
+ "grad_norm": 5.5625,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.745,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6704707560627675,
730
+ "grad_norm": 4.96875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.82,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.6776034236804565,
737
+ "grad_norm": 4.78125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8025,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.6847360912981455,
744
+ "grad_norm": 5.90625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.84,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6918687589158345,
751
+ "grad_norm": 4.21875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.8552,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6990014265335235,
758
+ "grad_norm": 4.15625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.8082,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7061340941512125,
765
+ "grad_norm": 4.71875,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8265,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7132667617689016,
772
+ "grad_norm": 4.875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8369,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7132667617689016,
779
+ "eval/nano_beir.arguana": 0.47952715537701884,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7132667617689016,
784
+ "eval/nano_beir.climate_fever": 0.22702207934204544,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7132667617689016,
789
+ "eval/nano_beir.dbpedia_entity": 0.3287396364958694,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7132667617689016,
794
+ "eval/nano_beir.fever": 0.5532628324042129,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7132667617689016,
799
+ "eval/nano_beir.fiqa": 0.4160855038689836,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7132667617689016,
804
+ "eval/nano_beir.hotpotqa": 0.6599716759215029,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7132667617689016,
809
+ "eval/nano_beir.nfcorpus": 0.3449425020579428,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7132667617689016,
814
+ "eval/nano_beir.nq": 0.39316532973343116,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7132667617689016,
819
+ "eval/nano_beir.quora": 0.8295275476072758,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7132667617689016,
824
+ "eval/nano_beir.scidocs": 0.2526832556333601,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7132667617689016,
829
+ "eval/nano_beir.scifact": 0.7219633063518482,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7132667617689016,
834
+ "eval/nano_beir.webis_touche2020": 0.2787587910246796,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7132667617689016,
839
+ "eval/avg": 0.4571374679848476,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7203994293865906,
844
+ "grad_norm": 5.15625,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8872,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7275320970042796,
851
+ "grad_norm": 5.65625,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8334,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7346647646219686,
858
+ "grad_norm": 5.1875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8014,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7417974322396577,
865
+ "grad_norm": 4.9375,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.848,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7489300998573466,
872
+ "grad_norm": 4.40625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.7203,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7560627674750356,
879
+ "grad_norm": 5.0625,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8405,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7631954350927247,
886
+ "grad_norm": 4.5625,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.8418,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7703281027104137,
893
+ "grad_norm": 5.75,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.8853,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7774607703281027,
900
+ "grad_norm": 4.09375,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.7733,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7845934379457917,
907
+ "grad_norm": 4.40625,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.7698,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7917261055634808,
914
+ "grad_norm": 4.84375,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.7759,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7988587731811697,
921
+ "grad_norm": 5.90625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8025,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8059914407988588,
928
+ "grad_norm": 5.53125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.8364,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8131241084165478,
935
+ "grad_norm": 5.03125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.7523,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.8202567760342369,
942
+ "grad_norm": 5.46875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7333,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.8273894436519258,
949
+ "grad_norm": 4.875,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.7325,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8345221112696148,
956
+ "grad_norm": 5.0,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7996,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8416547788873039,
963
+ "grad_norm": 4.6875,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7946,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8487874465049928,
970
+ "grad_norm": 5.21875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.9133,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8559201141226819,
977
+ "grad_norm": 4.78125,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.7218,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8630527817403709,
984
+ "grad_norm": 4.28125,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8453,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8701854493580599,
991
+ "grad_norm": 5.1875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.7261,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8773181169757489,
998
+ "grad_norm": 3.828125,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.6459,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.884450784593438,
1005
+ "grad_norm": 5.3125,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.7202,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.891583452211127,
1012
+ "grad_norm": 5.90625,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.7874,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8987161198288159,
1019
+ "grad_norm": 4.09375,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8524,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.905848787446505,
1026
+ "grad_norm": 5.75,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.7854,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.912981455064194,
1033
+ "grad_norm": 5.3125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8386,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.920114122681883,
1040
+ "grad_norm": 4.71875,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.8279,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.927246790299572,
1047
+ "grad_norm": 4.71875,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.8345,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9343794579172611,
1054
+ "grad_norm": 5.15625,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7596,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9415121255349501,
1061
+ "grad_norm": 4.4375,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.6704,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.948644793152639,
1068
+ "grad_norm": 5.6875,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7792,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.9557774607703281,
1075
+ "grad_norm": 4.78125,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.813,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.9629101283880172,
1082
+ "grad_norm": 4.4375,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.834,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9700427960057061,
1089
+ "grad_norm": 4.1875,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.8299,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9771754636233951,
1096
+ "grad_norm": 4.3125,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.7407,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9843081312410842,
1103
+ "grad_norm": 5.0,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.8821,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9914407988587732,
1110
+ "grad_norm": 4.71875,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8595,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9985734664764622,
1117
+ "grad_norm": 4.5,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.869,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.005706134094151,
1124
+ "grad_norm": 4.28125,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.8397,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0128388017118402,
1131
+ "grad_norm": 5.34375,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8047,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0199714693295292,
1138
+ "grad_norm": 4.28125,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.72,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0271041369472182,
1145
+ "grad_norm": 5.59375,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.7724,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0342368045649073,
1152
+ "grad_norm": 4.3125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.8014,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0413694721825963,
1159
+ "grad_norm": 6.03125,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.8532,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0485021398002854,
1166
+ "grad_norm": 3.75,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.7285,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0556348074179742,
1173
+ "grad_norm": 4.5625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.8574,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0627674750356633,
1180
+ "grad_norm": 6.21875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.76,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0699001426533523,
1187
+ "grad_norm": 4.75,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.8441,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0770328102710414,
1194
+ "grad_norm": 4.5625,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.764,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0841654778887304,
1201
+ "grad_norm": 4.59375,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.7432,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.0912981455064195,
1208
+ "grad_norm": 4.1875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.7894,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0984308131241085,
1215
+ "grad_norm": 4.5,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.761,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.1055634807417973,
1222
+ "grad_norm": 6.03125,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.9432,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1126961483594864,
1229
+ "grad_norm": 4.1875,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7863,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1198288159771754,
1236
+ "grad_norm": 4.21875,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.7732,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1269614835948645,
1243
+ "grad_norm": 4.65625,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.7651,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1340941512125535,
1250
+ "grad_norm": 4.34375,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8021,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1412268188302426,
1257
+ "grad_norm": 5.0625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8561,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1483594864479316,
1264
+ "grad_norm": 5.03125,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7551,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1554921540656204,
1271
+ "grad_norm": 5.375,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.8045,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1626248216833095,
1278
+ "grad_norm": 4.21875,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.7145,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1697574893009985,
1285
+ "grad_norm": 4.125,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8007,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1768901569186876,
1292
+ "grad_norm": 5.03125,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8009,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1840228245363766,
1299
+ "grad_norm": 4.84375,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.833,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1911554921540657,
1306
+ "grad_norm": 5.21875,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8149,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1982881597717547,
1313
+ "grad_norm": 4.1875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.808,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2054208273894436,
1320
+ "grad_norm": 4.625,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7549,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2125534950071326,
1327
+ "grad_norm": 5.71875,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7755,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.2196861626248217,
1334
+ "grad_norm": 4.5625,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8038,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.2268188302425107,
1341
+ "grad_norm": 4.09375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.7207,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.2339514978601998,
1348
+ "grad_norm": 4.90625,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7994,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.2410841654778888,
1355
+ "grad_norm": 4.3125,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7989,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2482168330955776,
1362
+ "grad_norm": 4.34375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.6516,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2553495007132667,
1369
+ "grad_norm": 4.59375,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7919,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2624821683309557,
1376
+ "grad_norm": 5.53125,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.9372,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2696148359486448,
1383
+ "grad_norm": 4.3125,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7683,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2767475035663338,
1390
+ "grad_norm": 4.3125,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.737,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2838801711840229,
1397
+ "grad_norm": 5.78125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.7507,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.291012838801712,
1404
+ "grad_norm": 4.875,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7016,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.298145506419401,
1411
+ "grad_norm": 4.5,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7151,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3052781740370898,
1418
+ "grad_norm": 4.5625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8559,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3124108416547788,
1425
+ "grad_norm": 4.15625,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7118,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3195435092724679,
1432
+ "grad_norm": 3.78125,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.7855,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.326676176890157,
1439
+ "grad_norm": 4.53125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.7438,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.333808844507846,
1446
+ "grad_norm": 4.65625,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.7018,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.340941512125535,
1453
+ "grad_norm": 6.9375,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.7083,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3480741797432239,
1460
+ "grad_norm": 4.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.8265,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.355206847360913,
1467
+ "grad_norm": 5.5625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.8735,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.362339514978602,
1474
+ "grad_norm": 5.71875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.7256,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.369472182596291,
1481
+ "grad_norm": 4.53125,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.7472,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.37660485021398,
1488
+ "grad_norm": 5.875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7601,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.383737517831669,
1495
+ "grad_norm": 5.75,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8857,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.3908701854493581,
1502
+ "grad_norm": 6.40625,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7296,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3980028530670472,
1509
+ "grad_norm": 4.40625,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7331,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.405135520684736,
1516
+ "grad_norm": 5.09375,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.8462,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.412268188302425,
1523
+ "grad_norm": 4.15625,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.7307,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4194008559201141,
1530
+ "grad_norm": 4.90625,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7711,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4265335235378032,
1537
+ "grad_norm": 5.78125,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.7485,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.4265335235378032,
1544
+ "eval/nano_beir.arguana": 0.47494654864753366,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.4265335235378032,
1549
+ "eval/nano_beir.climate_fever": 0.2496059194905162,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.4265335235378032,
1554
+ "eval/nano_beir.dbpedia_entity": 0.3371368383795953,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.4265335235378032,
1559
+ "eval/nano_beir.fever": 0.6065422117591313,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.4265335235378032,
1564
+ "eval/nano_beir.fiqa": 0.4135337243548038,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.4265335235378032,
1569
+ "eval/nano_beir.hotpotqa": 0.6804056480114672,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.4265335235378032,
1574
+ "eval/nano_beir.nfcorpus": 0.35238501477624823,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.4265335235378032,
1579
+ "eval/nano_beir.nq": 0.39770361480597743,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.4265335235378032,
1584
+ "eval/nano_beir.quora": 0.8217479889085385,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.4265335235378032,
1589
+ "eval/nano_beir.scidocs": 0.2592573629546752,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.4265335235378032,
1594
+ "eval/nano_beir.scifact": 0.7479618950231317,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.4265335235378032,
1599
+ "eval/nano_beir.webis_touche2020": 0.2719065772753103,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.4265335235378032,
1604
+ "eval/avg": 0.467761112032244,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.4336661911554922,
1609
+ "grad_norm": 3.71875,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.8505,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.440798858773181,
1616
+ "grad_norm": 6.15625,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.7449,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.44793152639087,
1623
+ "grad_norm": 4.0625,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.6961,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.4550641940085591,
1630
+ "grad_norm": 4.40625,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.8105,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.4621968616262482,
1637
+ "grad_norm": 5.59375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.8111,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.4693295292439372,
1644
+ "grad_norm": 4.625,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.7972,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.4764621968616263,
1651
+ "grad_norm": 3.953125,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.7745,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.4835948644793153,
1658
+ "grad_norm": 3.859375,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.7904,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.4907275320970044,
1665
+ "grad_norm": 4.875,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.7724,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.4978601997146934,
1672
+ "grad_norm": 5.59375,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.7693,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.5049928673323825,
1679
+ "grad_norm": 4.75,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.6929,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.5121255349500713,
1686
+ "grad_norm": 5.78125,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.8008,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.5192582025677603,
1693
+ "grad_norm": 6.28125,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.8399,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.5263908701854494,
1700
+ "grad_norm": 6.3125,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.7399,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.5335235378031382,
1707
+ "grad_norm": 4.1875,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.7796,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.5406562054208273,
1714
+ "grad_norm": 5.34375,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.8685,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.5477888730385163,
1721
+ "grad_norm": 4.8125,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.7448,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.5549215406562054,
1728
+ "grad_norm": 6.09375,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.8527,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.5620542082738944,
1735
+ "grad_norm": 5.15625,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.8704,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.5691868758915835,
1742
+ "grad_norm": 4.3125,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.7033,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.5763195435092725,
1749
+ "grad_norm": 5.6875,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.7664,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.5834522111269616,
1756
+ "grad_norm": 4.5625,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.7902,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.5905848787446506,
1763
+ "grad_norm": 4.6875,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.7585,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.5977175463623396,
1770
+ "grad_norm": 5.375,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.8124,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.6048502139800287,
1777
+ "grad_norm": 4.59375,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.7936,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.6119828815977175,
1784
+ "grad_norm": 3.5625,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.6785,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.6191155492154066,
1791
+ "grad_norm": 5.875,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.8321,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.6262482168330956,
1798
+ "grad_norm": 5.125,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.8184,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.6333808844507844,
1805
+ "grad_norm": 4.5,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.7187,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.6405135520684735,
1812
+ "grad_norm": 4.96875,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.6873,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.6476462196861625,
1819
+ "grad_norm": 4.96875,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.7537,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.6547788873038516,
1826
+ "grad_norm": 4.71875,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.8058,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.6619115549215406,
1833
+ "grad_norm": 5.375,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.8117,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.6690442225392297,
1840
+ "grad_norm": 5.15625,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.7938,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.6761768901569187,
1847
+ "grad_norm": 6.15625,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.7898,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.6833095577746078,
1854
+ "grad_norm": 4.03125,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.7896,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.6904422253922968,
1861
+ "grad_norm": 4.28125,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.6994,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.6975748930099859,
1868
+ "grad_norm": 5.71875,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.7416,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.7047075606276747,
1875
+ "grad_norm": 4.8125,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.7143,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.7118402282453637,
1882
+ "grad_norm": 5.28125,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.6699,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.7189728958630528,
1889
+ "grad_norm": 4.84375,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.7084,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.7261055634807418,
1896
+ "grad_norm": 5.125,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.7638,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.7332382310984307,
1903
+ "grad_norm": 5.0,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.7875,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.7403708987161197,
1910
+ "grad_norm": 3.921875,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.8864,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.7475035663338088,
1917
+ "grad_norm": 4.84375,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.7354,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.7546362339514978,
1924
+ "grad_norm": 3.90625,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.7581,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.7617689015691869,
1931
+ "grad_norm": 4.40625,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.738,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.768901569186876,
1938
+ "grad_norm": 4.53125,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.6686,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.776034236804565,
1945
+ "grad_norm": 5.125,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.6354,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.783166904422254,
1952
+ "grad_norm": 5.125,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.7869,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.790299572039943,
1959
+ "grad_norm": 4.21875,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.7463,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.797432239657632,
1966
+ "grad_norm": 3.9375,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.7524,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.804564907275321,
1973
+ "grad_norm": 6.15625,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.7558,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.81169757489301,
1980
+ "grad_norm": 4.40625,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.8002,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 1.818830242510699,
1987
+ "grad_norm": 5.5625,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.7829,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 1.825962910128388,
1994
+ "grad_norm": 4.375,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.7097,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 1.833095577746077,
2001
+ "grad_norm": 4.96875,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.6793,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 1.840228245363766,
2008
+ "grad_norm": 4.84375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.7081,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 1.847360912981455,
2015
+ "grad_norm": 4.25,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.911,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 1.854493580599144,
2022
+ "grad_norm": 5.5625,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.8128,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 1.861626248216833,
2029
+ "grad_norm": 4.34375,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.7784,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 1.8687589158345221,
2036
+ "grad_norm": 3.859375,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.6765,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 1.8758915834522112,
2043
+ "grad_norm": 4.96875,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.7464,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 1.8830242510699002,
2050
+ "grad_norm": 4.4375,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.7452,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 1.8901569186875893,
2057
+ "grad_norm": 4.40625,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.7475,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 1.8972895863052783,
2064
+ "grad_norm": 3.6875,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.761,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 1.9044222539229672,
2071
+ "grad_norm": 4.5625,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.7183,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 1.9115549215406562,
2078
+ "grad_norm": 4.875,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.756,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 1.9186875891583453,
2085
+ "grad_norm": 4.8125,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.6997,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 1.925820256776034,
2092
+ "grad_norm": 4.75,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.7978,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 1.9329529243937231,
2099
+ "grad_norm": 4.5,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.7099,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 1.9400855920114122,
2106
+ "grad_norm": 4.8125,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.7029,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 1.9472182596291012,
2113
+ "grad_norm": 5.875,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.7328,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 1.9543509272467903,
2120
+ "grad_norm": 3.875,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.7982,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 1.9614835948644793,
2127
+ "grad_norm": 4.71875,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.7126,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 1.9686162624821684,
2134
+ "grad_norm": 4.125,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.6598,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 1.9757489300998574,
2141
+ "grad_norm": 4.03125,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.7039,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 1.9828815977175465,
2148
+ "grad_norm": 4.78125,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.6858,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 1.9900142653352355,
2155
+ "grad_norm": 4.53125,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.7924,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 1.9971469329529246,
2162
+ "grad_norm": 5.40625,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.7606,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.0042796005706136,
2169
+ "grad_norm": 5.0,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.7584,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.011412268188302,
2176
+ "grad_norm": 5.34375,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.6966,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.0185449358059913,
2183
+ "grad_norm": 4.5625,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.8236,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.0256776034236803,
2190
+ "grad_norm": 4.1875,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.7729,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.0328102710413694,
2197
+ "grad_norm": 4.875,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6784,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.0399429386590584,
2204
+ "grad_norm": 8.8125,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.8262,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.0470756062767475,
2211
+ "grad_norm": 4.15625,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.7451,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.0542082738944365,
2218
+ "grad_norm": 5.59375,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.7622,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.0613409415121255,
2225
+ "grad_norm": 5.9375,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.6851,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.0684736091298146,
2232
+ "grad_norm": 4.375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.7706,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.0756062767475036,
2239
+ "grad_norm": 4.84375,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.7432,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.0827389443651927,
2246
+ "grad_norm": 4.21875,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.7158,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.0898716119828817,
2253
+ "grad_norm": 4.375,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.7035,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.097004279600571,
2260
+ "grad_norm": 4.65625,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.7611,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.10413694721826,
2267
+ "grad_norm": 4.625,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.6861,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.1112696148359484,
2274
+ "grad_norm": 5.875,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.8394,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.1184022824536375,
2281
+ "grad_norm": 4.1875,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.6647,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.1255349500713265,
2288
+ "grad_norm": 4.3125,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.7305,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.1326676176890156,
2295
+ "grad_norm": 5.59375,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.7693,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.1398002853067046,
2302
+ "grad_norm": 5.28125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.7392,
2305
+ "step": 3000
2306
+ }
2307
+ ],
2308
+ "logging_steps": 10,
2309
+ "max_steps": 5000,
2310
+ "num_input_tokens_seen": 0,
2311
+ "num_train_epochs": 4,
2312
+ "save_steps": 1000,
2313
+ "stateful_callbacks": {
2314
+ "TrainerControl": {
2315
+ "args": {
2316
+ "should_epoch_stop": false,
2317
+ "should_evaluate": false,
2318
+ "should_log": false,
2319
+ "should_save": true,
2320
+ "should_training_stop": false
2321
+ },
2322
+ "attributes": {}
2323
+ }
2324
+ },
2325
+ "total_flos": 0.0,
2326
+ "train_batch_size": 16,
2327
+ "trial_name": null,
2328
+ "trial_params": null
2329
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-3000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af7cf7c3c741add5e8996639d61d87fe74bb40ad0faa1b29901130f4b3a60fc3
3
+ size 6353
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86aa8a0a858b4170a0a0701c1871f2d27d9c905bd960a613015f68b2f98fbb58
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcd48107e84feb43b1ddcf1ad31643963a6e81e3b99f3a74da919de2742d9444
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac0e770c61e3acef7bd144efee65a450ed487b244ed293a8dd801394ffcae775
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fb125336725f7741cb4daa1e3d06e225bbacfde8d41c4dcabb6762c222e62c6
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:153c865f77c7129ba565bded50f334683d51c80f20e3cfec39e62f8737b86f0d
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d017ce00fcebac7edc058ddd138f194eb0340f2d8ad0879bdab08f922ed0846e
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9218a8f161ba45a13a24f24cd020f32b0abec4bcc53a2764e48b480568dd2d13
3
+ size 1465
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/trainer_state.json ADDED
@@ -0,0 +1,3094 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.8530670470756063,
6
+ "eval_steps": 1000.0,
7
+ "global_step": 4000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0,
14
+ "eval/nano_beir.arguana": 0.5201586889046041,
15
+ "step": 0
16
+ },
17
+ {
18
+ "epoch": 0,
19
+ "eval/nano_beir.climate_fever": 0.23997409265084552,
20
+ "step": 0
21
+ },
22
+ {
23
+ "epoch": 0,
24
+ "eval/nano_beir.dbpedia_entity": 0.34286887314902237,
25
+ "step": 0
26
+ },
27
+ {
28
+ "epoch": 0,
29
+ "eval/nano_beir.fever": 0.6213496755517833,
30
+ "step": 0
31
+ },
32
+ {
33
+ "epoch": 0,
34
+ "eval/nano_beir.fiqa": 0.4542903465876373,
35
+ "step": 0
36
+ },
37
+ {
38
+ "epoch": 0,
39
+ "eval/nano_beir.hotpotqa": 0.6560051657630788,
40
+ "step": 0
41
+ },
42
+ {
43
+ "epoch": 0,
44
+ "eval/nano_beir.nfcorpus": 0.3328835143785305,
45
+ "step": 0
46
+ },
47
+ {
48
+ "epoch": 0,
49
+ "eval/nano_beir.nq": 0.4570725655218753,
50
+ "step": 0
51
+ },
52
+ {
53
+ "epoch": 0,
54
+ "eval/nano_beir.quora": 0.832098944676056,
55
+ "step": 0
56
+ },
57
+ {
58
+ "epoch": 0,
59
+ "eval/nano_beir.scidocs": 0.25908236715595256,
60
+ "step": 0
61
+ },
62
+ {
63
+ "epoch": 0,
64
+ "eval/nano_beir.scifact": 0.7391066781087818,
65
+ "step": 0
66
+ },
67
+ {
68
+ "epoch": 0,
69
+ "eval/nano_beir.webis_touche2020": 0.30431141846717047,
70
+ "step": 0
71
+ },
72
+ {
73
+ "epoch": 0,
74
+ "eval/avg": 0.4799335275762782,
75
+ "step": 0
76
+ },
77
+ {
78
+ "epoch": 0.007132667617689016,
79
+ "grad_norm": 18.375,
80
+ "learning_rate": 1.8e-06,
81
+ "loss": 1.6901,
82
+ "step": 10
83
+ },
84
+ {
85
+ "epoch": 0.014265335235378032,
86
+ "grad_norm": 10.75,
87
+ "learning_rate": 3.8e-06,
88
+ "loss": 1.7675,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.021398002853067047,
93
+ "grad_norm": 21.0,
94
+ "learning_rate": 5.8e-06,
95
+ "loss": 1.527,
96
+ "step": 30
97
+ },
98
+ {
99
+ "epoch": 0.028530670470756064,
100
+ "grad_norm": 19.0,
101
+ "learning_rate": 7.8e-06,
102
+ "loss": 1.7708,
103
+ "step": 40
104
+ },
105
+ {
106
+ "epoch": 0.03566333808844508,
107
+ "grad_norm": 17.125,
108
+ "learning_rate": 9.800000000000001e-06,
109
+ "loss": 1.4093,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 0.042796005706134094,
114
+ "grad_norm": 15.625,
115
+ "learning_rate": 1.18e-05,
116
+ "loss": 1.579,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.04992867332382311,
121
+ "grad_norm": 14.0,
122
+ "learning_rate": 1.3800000000000002e-05,
123
+ "loss": 1.4845,
124
+ "step": 70
125
+ },
126
+ {
127
+ "epoch": 0.05706134094151213,
128
+ "grad_norm": 13.6875,
129
+ "learning_rate": 1.58e-05,
130
+ "loss": 1.598,
131
+ "step": 80
132
+ },
133
+ {
134
+ "epoch": 0.06419400855920114,
135
+ "grad_norm": 15.5625,
136
+ "learning_rate": 1.78e-05,
137
+ "loss": 1.3226,
138
+ "step": 90
139
+ },
140
+ {
141
+ "epoch": 0.07132667617689016,
142
+ "grad_norm": 13.8125,
143
+ "learning_rate": 1.9800000000000004e-05,
144
+ "loss": 1.4969,
145
+ "step": 100
146
+ },
147
+ {
148
+ "epoch": 0.07845934379457917,
149
+ "grad_norm": 13.125,
150
+ "learning_rate": 2.18e-05,
151
+ "loss": 1.3867,
152
+ "step": 110
153
+ },
154
+ {
155
+ "epoch": 0.08559201141226819,
156
+ "grad_norm": 14.5625,
157
+ "learning_rate": 2.38e-05,
158
+ "loss": 1.1904,
159
+ "step": 120
160
+ },
161
+ {
162
+ "epoch": 0.09272467902995721,
163
+ "grad_norm": 9.5625,
164
+ "learning_rate": 2.58e-05,
165
+ "loss": 1.1668,
166
+ "step": 130
167
+ },
168
+ {
169
+ "epoch": 0.09985734664764621,
170
+ "grad_norm": 8.1875,
171
+ "learning_rate": 2.7800000000000005e-05,
172
+ "loss": 1.124,
173
+ "step": 140
174
+ },
175
+ {
176
+ "epoch": 0.10699001426533523,
177
+ "grad_norm": 9.75,
178
+ "learning_rate": 2.98e-05,
179
+ "loss": 1.1571,
180
+ "step": 150
181
+ },
182
+ {
183
+ "epoch": 0.11412268188302425,
184
+ "grad_norm": 12.25,
185
+ "learning_rate": 3.18e-05,
186
+ "loss": 1.2145,
187
+ "step": 160
188
+ },
189
+ {
190
+ "epoch": 0.12125534950071326,
191
+ "grad_norm": 10.625,
192
+ "learning_rate": 3.38e-05,
193
+ "loss": 1.155,
194
+ "step": 170
195
+ },
196
+ {
197
+ "epoch": 0.12838801711840228,
198
+ "grad_norm": 8.5,
199
+ "learning_rate": 3.58e-05,
200
+ "loss": 1.1513,
201
+ "step": 180
202
+ },
203
+ {
204
+ "epoch": 0.1355206847360913,
205
+ "grad_norm": 9.5,
206
+ "learning_rate": 3.7800000000000004e-05,
207
+ "loss": 0.9803,
208
+ "step": 190
209
+ },
210
+ {
211
+ "epoch": 0.14265335235378032,
212
+ "grad_norm": 8.0625,
213
+ "learning_rate": 3.9800000000000005e-05,
214
+ "loss": 1.0302,
215
+ "step": 200
216
+ },
217
+ {
218
+ "epoch": 0.14978601997146934,
219
+ "grad_norm": 9.25,
220
+ "learning_rate": 4.18e-05,
221
+ "loss": 0.9878,
222
+ "step": 210
223
+ },
224
+ {
225
+ "epoch": 0.15691868758915833,
226
+ "grad_norm": 10.25,
227
+ "learning_rate": 4.38e-05,
228
+ "loss": 0.9816,
229
+ "step": 220
230
+ },
231
+ {
232
+ "epoch": 0.16405135520684735,
233
+ "grad_norm": 8.6875,
234
+ "learning_rate": 4.58e-05,
235
+ "loss": 1.0415,
236
+ "step": 230
237
+ },
238
+ {
239
+ "epoch": 0.17118402282453637,
240
+ "grad_norm": 10.1875,
241
+ "learning_rate": 4.78e-05,
242
+ "loss": 1.1126,
243
+ "step": 240
244
+ },
245
+ {
246
+ "epoch": 0.1783166904422254,
247
+ "grad_norm": 7.78125,
248
+ "learning_rate": 4.9800000000000004e-05,
249
+ "loss": 1.0127,
250
+ "step": 250
251
+ },
252
+ {
253
+ "epoch": 0.18544935805991442,
254
+ "grad_norm": 5.71875,
255
+ "learning_rate": 5.1800000000000005e-05,
256
+ "loss": 0.9577,
257
+ "step": 260
258
+ },
259
+ {
260
+ "epoch": 0.19258202567760344,
261
+ "grad_norm": 6.21875,
262
+ "learning_rate": 5.380000000000001e-05,
263
+ "loss": 0.9518,
264
+ "step": 270
265
+ },
266
+ {
267
+ "epoch": 0.19971469329529243,
268
+ "grad_norm": 6.28125,
269
+ "learning_rate": 5.580000000000001e-05,
270
+ "loss": 0.7952,
271
+ "step": 280
272
+ },
273
+ {
274
+ "epoch": 0.20684736091298145,
275
+ "grad_norm": 6.875,
276
+ "learning_rate": 5.7799999999999995e-05,
277
+ "loss": 0.9201,
278
+ "step": 290
279
+ },
280
+ {
281
+ "epoch": 0.21398002853067047,
282
+ "grad_norm": 6.09375,
283
+ "learning_rate": 5.9800000000000003e-05,
284
+ "loss": 0.951,
285
+ "step": 300
286
+ },
287
+ {
288
+ "epoch": 0.2211126961483595,
289
+ "grad_norm": 7.65625,
290
+ "learning_rate": 6.18e-05,
291
+ "loss": 0.9199,
292
+ "step": 310
293
+ },
294
+ {
295
+ "epoch": 0.2282453637660485,
296
+ "grad_norm": 7.3125,
297
+ "learning_rate": 6.38e-05,
298
+ "loss": 0.8773,
299
+ "step": 320
300
+ },
301
+ {
302
+ "epoch": 0.23537803138373753,
303
+ "grad_norm": 4.875,
304
+ "learning_rate": 6.58e-05,
305
+ "loss": 0.9386,
306
+ "step": 330
307
+ },
308
+ {
309
+ "epoch": 0.24251069900142652,
310
+ "grad_norm": 5.1875,
311
+ "learning_rate": 6.780000000000001e-05,
312
+ "loss": 0.9364,
313
+ "step": 340
314
+ },
315
+ {
316
+ "epoch": 0.24964336661911554,
317
+ "grad_norm": 7.21875,
318
+ "learning_rate": 6.98e-05,
319
+ "loss": 0.9053,
320
+ "step": 350
321
+ },
322
+ {
323
+ "epoch": 0.25677603423680456,
324
+ "grad_norm": 5.8125,
325
+ "learning_rate": 7.18e-05,
326
+ "loss": 0.8493,
327
+ "step": 360
328
+ },
329
+ {
330
+ "epoch": 0.26390870185449355,
331
+ "grad_norm": 7.1875,
332
+ "learning_rate": 7.38e-05,
333
+ "loss": 0.8218,
334
+ "step": 370
335
+ },
336
+ {
337
+ "epoch": 0.2710413694721826,
338
+ "grad_norm": 5.625,
339
+ "learning_rate": 7.58e-05,
340
+ "loss": 0.8975,
341
+ "step": 380
342
+ },
343
+ {
344
+ "epoch": 0.2781740370898716,
345
+ "grad_norm": 5.1875,
346
+ "learning_rate": 7.780000000000001e-05,
347
+ "loss": 0.8891,
348
+ "step": 390
349
+ },
350
+ {
351
+ "epoch": 0.28530670470756064,
352
+ "grad_norm": 5.78125,
353
+ "learning_rate": 7.98e-05,
354
+ "loss": 0.7831,
355
+ "step": 400
356
+ },
357
+ {
358
+ "epoch": 0.29243937232524964,
359
+ "grad_norm": 5.125,
360
+ "learning_rate": 8.18e-05,
361
+ "loss": 0.9017,
362
+ "step": 410
363
+ },
364
+ {
365
+ "epoch": 0.2995720399429387,
366
+ "grad_norm": 4.96875,
367
+ "learning_rate": 8.38e-05,
368
+ "loss": 0.7839,
369
+ "step": 420
370
+ },
371
+ {
372
+ "epoch": 0.3067047075606277,
373
+ "grad_norm": 5.5625,
374
+ "learning_rate": 8.58e-05,
375
+ "loss": 0.886,
376
+ "step": 430
377
+ },
378
+ {
379
+ "epoch": 0.31383737517831667,
380
+ "grad_norm": 6.1875,
381
+ "learning_rate": 8.78e-05,
382
+ "loss": 0.863,
383
+ "step": 440
384
+ },
385
+ {
386
+ "epoch": 0.3209700427960057,
387
+ "grad_norm": 4.625,
388
+ "learning_rate": 8.98e-05,
389
+ "loss": 0.884,
390
+ "step": 450
391
+ },
392
+ {
393
+ "epoch": 0.3281027104136947,
394
+ "grad_norm": 5.71875,
395
+ "learning_rate": 9.180000000000001e-05,
396
+ "loss": 0.9209,
397
+ "step": 460
398
+ },
399
+ {
400
+ "epoch": 0.33523537803138376,
401
+ "grad_norm": 6.09375,
402
+ "learning_rate": 9.38e-05,
403
+ "loss": 0.8775,
404
+ "step": 470
405
+ },
406
+ {
407
+ "epoch": 0.34236804564907275,
408
+ "grad_norm": 5.8125,
409
+ "learning_rate": 9.58e-05,
410
+ "loss": 0.7956,
411
+ "step": 480
412
+ },
413
+ {
414
+ "epoch": 0.34950071326676174,
415
+ "grad_norm": 5.71875,
416
+ "learning_rate": 9.78e-05,
417
+ "loss": 0.8632,
418
+ "step": 490
419
+ },
420
+ {
421
+ "epoch": 0.3566333808844508,
422
+ "grad_norm": 4.4375,
423
+ "learning_rate": 9.98e-05,
424
+ "loss": 0.8308,
425
+ "step": 500
426
+ },
427
+ {
428
+ "epoch": 0.3637660485021398,
429
+ "grad_norm": 5.25,
430
+ "learning_rate": 9.999901304280685e-05,
431
+ "loss": 0.884,
432
+ "step": 510
433
+ },
434
+ {
435
+ "epoch": 0.37089871611982883,
436
+ "grad_norm": 5.28125,
437
+ "learning_rate": 9.999560138895238e-05,
438
+ "loss": 0.8663,
439
+ "step": 520
440
+ },
441
+ {
442
+ "epoch": 0.3780313837375178,
443
+ "grad_norm": 5.25,
444
+ "learning_rate": 9.99897530200195e-05,
445
+ "loss": 0.8722,
446
+ "step": 530
447
+ },
448
+ {
449
+ "epoch": 0.38516405135520687,
450
+ "grad_norm": 4.6875,
451
+ "learning_rate": 9.998146822104943e-05,
452
+ "loss": 0.8391,
453
+ "step": 540
454
+ },
455
+ {
456
+ "epoch": 0.39229671897289586,
457
+ "grad_norm": 5.5625,
458
+ "learning_rate": 9.997074739583162e-05,
459
+ "loss": 0.7823,
460
+ "step": 550
461
+ },
462
+ {
463
+ "epoch": 0.39942938659058486,
464
+ "grad_norm": 4.875,
465
+ "learning_rate": 9.995759106688393e-05,
466
+ "loss": 0.8606,
467
+ "step": 560
468
+ },
469
+ {
470
+ "epoch": 0.4065620542082739,
471
+ "grad_norm": 5.625,
472
+ "learning_rate": 9.99419998754273e-05,
473
+ "loss": 0.8719,
474
+ "step": 570
475
+ },
476
+ {
477
+ "epoch": 0.4136947218259629,
478
+ "grad_norm": 4.71875,
479
+ "learning_rate": 9.992397458135438e-05,
480
+ "loss": 0.8619,
481
+ "step": 580
482
+ },
483
+ {
484
+ "epoch": 0.42082738944365194,
485
+ "grad_norm": 5.0,
486
+ "learning_rate": 9.990351606319261e-05,
487
+ "loss": 0.8439,
488
+ "step": 590
489
+ },
490
+ {
491
+ "epoch": 0.42796005706134094,
492
+ "grad_norm": 5.0625,
493
+ "learning_rate": 9.988062531806126e-05,
494
+ "loss": 0.886,
495
+ "step": 600
496
+ },
497
+ {
498
+ "epoch": 0.43509272467902993,
499
+ "grad_norm": 5.125,
500
+ "learning_rate": 9.985530346162299e-05,
501
+ "loss": 0.8346,
502
+ "step": 610
503
+ },
504
+ {
505
+ "epoch": 0.442225392296719,
506
+ "grad_norm": 5.40625,
507
+ "learning_rate": 9.982755172802933e-05,
508
+ "loss": 0.8145,
509
+ "step": 620
510
+ },
511
+ {
512
+ "epoch": 0.44935805991440797,
513
+ "grad_norm": 5.0625,
514
+ "learning_rate": 9.979737146986064e-05,
515
+ "loss": 0.8739,
516
+ "step": 630
517
+ },
518
+ {
519
+ "epoch": 0.456490727532097,
520
+ "grad_norm": 4.75,
521
+ "learning_rate": 9.976476415806013e-05,
522
+ "loss": 0.7826,
523
+ "step": 640
524
+ },
525
+ {
526
+ "epoch": 0.463623395149786,
527
+ "grad_norm": 5.6875,
528
+ "learning_rate": 9.972973138186217e-05,
529
+ "loss": 0.8541,
530
+ "step": 650
531
+ },
532
+ {
533
+ "epoch": 0.47075606276747506,
534
+ "grad_norm": 4.3125,
535
+ "learning_rate": 9.969227484871484e-05,
536
+ "loss": 0.8208,
537
+ "step": 660
538
+ },
539
+ {
540
+ "epoch": 0.47788873038516405,
541
+ "grad_norm": 5.375,
542
+ "learning_rate": 9.965239638419672e-05,
543
+ "loss": 0.7761,
544
+ "step": 670
545
+ },
546
+ {
547
+ "epoch": 0.48502139800285304,
548
+ "grad_norm": 5.15625,
549
+ "learning_rate": 9.961009793192792e-05,
550
+ "loss": 0.8346,
551
+ "step": 680
552
+ },
553
+ {
554
+ "epoch": 0.4921540656205421,
555
+ "grad_norm": 5.03125,
556
+ "learning_rate": 9.956538155347534e-05,
557
+ "loss": 0.9192,
558
+ "step": 690
559
+ },
560
+ {
561
+ "epoch": 0.4992867332382311,
562
+ "grad_norm": 4.96875,
563
+ "learning_rate": 9.951824942825215e-05,
564
+ "loss": 0.8037,
565
+ "step": 700
566
+ },
567
+ {
568
+ "epoch": 0.5064194008559201,
569
+ "grad_norm": 4.53125,
570
+ "learning_rate": 9.946870385341167e-05,
571
+ "loss": 0.7669,
572
+ "step": 710
573
+ },
574
+ {
575
+ "epoch": 0.5135520684736091,
576
+ "grad_norm": 6.875,
577
+ "learning_rate": 9.941674724373531e-05,
578
+ "loss": 0.7915,
579
+ "step": 720
580
+ },
581
+ {
582
+ "epoch": 0.5206847360912982,
583
+ "grad_norm": 5.0625,
584
+ "learning_rate": 9.936238213151491e-05,
585
+ "loss": 0.7863,
586
+ "step": 730
587
+ },
588
+ {
589
+ "epoch": 0.5278174037089871,
590
+ "grad_norm": 5.21875,
591
+ "learning_rate": 9.930561116642935e-05,
592
+ "loss": 0.7789,
593
+ "step": 740
594
+ },
595
+ {
596
+ "epoch": 0.5349500713266762,
597
+ "grad_norm": 4.0,
598
+ "learning_rate": 9.924643711541539e-05,
599
+ "loss": 0.7759,
600
+ "step": 750
601
+ },
602
+ {
603
+ "epoch": 0.5420827389443652,
604
+ "grad_norm": 5.34375,
605
+ "learning_rate": 9.918486286253279e-05,
606
+ "loss": 0.8504,
607
+ "step": 760
608
+ },
609
+ {
610
+ "epoch": 0.5492154065620543,
611
+ "grad_norm": 4.59375,
612
+ "learning_rate": 9.912089140882376e-05,
613
+ "loss": 0.809,
614
+ "step": 770
615
+ },
616
+ {
617
+ "epoch": 0.5563480741797432,
618
+ "grad_norm": 5.53125,
619
+ "learning_rate": 9.90545258721667e-05,
620
+ "loss": 0.7721,
621
+ "step": 780
622
+ },
623
+ {
624
+ "epoch": 0.5634807417974322,
625
+ "grad_norm": 4.8125,
626
+ "learning_rate": 9.898576948712426e-05,
627
+ "loss": 0.7319,
628
+ "step": 790
629
+ },
630
+ {
631
+ "epoch": 0.5706134094151213,
632
+ "grad_norm": 4.84375,
633
+ "learning_rate": 9.891462560478562e-05,
634
+ "loss": 0.7726,
635
+ "step": 800
636
+ },
637
+ {
638
+ "epoch": 0.5777460770328102,
639
+ "grad_norm": 4.625,
640
+ "learning_rate": 9.884109769260325e-05,
641
+ "loss": 0.8522,
642
+ "step": 810
643
+ },
644
+ {
645
+ "epoch": 0.5848787446504993,
646
+ "grad_norm": 4.34375,
647
+ "learning_rate": 9.876518933422386e-05,
648
+ "loss": 0.7761,
649
+ "step": 820
650
+ },
651
+ {
652
+ "epoch": 0.5920114122681883,
653
+ "grad_norm": 5.0625,
654
+ "learning_rate": 9.868690422931372e-05,
655
+ "loss": 0.8702,
656
+ "step": 830
657
+ },
658
+ {
659
+ "epoch": 0.5991440798858774,
660
+ "grad_norm": 5.25,
661
+ "learning_rate": 9.860624619337844e-05,
662
+ "loss": 0.8636,
663
+ "step": 840
664
+ },
665
+ {
666
+ "epoch": 0.6062767475035663,
667
+ "grad_norm": 3.96875,
668
+ "learning_rate": 9.852321915757687e-05,
669
+ "loss": 0.6974,
670
+ "step": 850
671
+ },
672
+ {
673
+ "epoch": 0.6134094151212554,
674
+ "grad_norm": 5.03125,
675
+ "learning_rate": 9.843782716852963e-05,
676
+ "loss": 0.8017,
677
+ "step": 860
678
+ },
679
+ {
680
+ "epoch": 0.6205420827389444,
681
+ "grad_norm": 5.40625,
682
+ "learning_rate": 9.835007438812177e-05,
683
+ "loss": 0.7799,
684
+ "step": 870
685
+ },
686
+ {
687
+ "epoch": 0.6276747503566333,
688
+ "grad_norm": 5.125,
689
+ "learning_rate": 9.82599650933e-05,
690
+ "loss": 0.7808,
691
+ "step": 880
692
+ },
693
+ {
694
+ "epoch": 0.6348074179743224,
695
+ "grad_norm": 5.4375,
696
+ "learning_rate": 9.816750367586425e-05,
697
+ "loss": 0.8337,
698
+ "step": 890
699
+ },
700
+ {
701
+ "epoch": 0.6419400855920114,
702
+ "grad_norm": 4.90625,
703
+ "learning_rate": 9.807269464225355e-05,
704
+ "loss": 0.8043,
705
+ "step": 900
706
+ },
707
+ {
708
+ "epoch": 0.6490727532097005,
709
+ "grad_norm": 4.59375,
710
+ "learning_rate": 9.797554261332645e-05,
711
+ "loss": 0.8328,
712
+ "step": 910
713
+ },
714
+ {
715
+ "epoch": 0.6562054208273894,
716
+ "grad_norm": 4.59375,
717
+ "learning_rate": 9.787605232413574e-05,
718
+ "loss": 0.7745,
719
+ "step": 920
720
+ },
721
+ {
722
+ "epoch": 0.6633380884450785,
723
+ "grad_norm": 5.5625,
724
+ "learning_rate": 9.777422862369783e-05,
725
+ "loss": 0.745,
726
+ "step": 930
727
+ },
728
+ {
729
+ "epoch": 0.6704707560627675,
730
+ "grad_norm": 4.96875,
731
+ "learning_rate": 9.767007647475617e-05,
732
+ "loss": 0.82,
733
+ "step": 940
734
+ },
735
+ {
736
+ "epoch": 0.6776034236804565,
737
+ "grad_norm": 4.78125,
738
+ "learning_rate": 9.756360095353956e-05,
739
+ "loss": 0.8025,
740
+ "step": 950
741
+ },
742
+ {
743
+ "epoch": 0.6847360912981455,
744
+ "grad_norm": 5.90625,
745
+ "learning_rate": 9.745480724951473e-05,
746
+ "loss": 0.84,
747
+ "step": 960
748
+ },
749
+ {
750
+ "epoch": 0.6918687589158345,
751
+ "grad_norm": 4.21875,
752
+ "learning_rate": 9.73437006651333e-05,
753
+ "loss": 0.8552,
754
+ "step": 970
755
+ },
756
+ {
757
+ "epoch": 0.6990014265335235,
758
+ "grad_norm": 4.15625,
759
+ "learning_rate": 9.723028661557345e-05,
760
+ "loss": 0.8082,
761
+ "step": 980
762
+ },
763
+ {
764
+ "epoch": 0.7061340941512125,
765
+ "grad_norm": 4.71875,
766
+ "learning_rate": 9.711457062847595e-05,
767
+ "loss": 0.8265,
768
+ "step": 990
769
+ },
770
+ {
771
+ "epoch": 0.7132667617689016,
772
+ "grad_norm": 4.875,
773
+ "learning_rate": 9.699655834367478e-05,
774
+ "loss": 0.8369,
775
+ "step": 1000
776
+ },
777
+ {
778
+ "epoch": 0.7132667617689016,
779
+ "eval/nano_beir.arguana": 0.47952715537701884,
780
+ "step": 1000
781
+ },
782
+ {
783
+ "epoch": 0.7132667617689016,
784
+ "eval/nano_beir.climate_fever": 0.22702207934204544,
785
+ "step": 1000
786
+ },
787
+ {
788
+ "epoch": 0.7132667617689016,
789
+ "eval/nano_beir.dbpedia_entity": 0.3287396364958694,
790
+ "step": 1000
791
+ },
792
+ {
793
+ "epoch": 0.7132667617689016,
794
+ "eval/nano_beir.fever": 0.5532628324042129,
795
+ "step": 1000
796
+ },
797
+ {
798
+ "epoch": 0.7132667617689016,
799
+ "eval/nano_beir.fiqa": 0.4160855038689836,
800
+ "step": 1000
801
+ },
802
+ {
803
+ "epoch": 0.7132667617689016,
804
+ "eval/nano_beir.hotpotqa": 0.6599716759215029,
805
+ "step": 1000
806
+ },
807
+ {
808
+ "epoch": 0.7132667617689016,
809
+ "eval/nano_beir.nfcorpus": 0.3449425020579428,
810
+ "step": 1000
811
+ },
812
+ {
813
+ "epoch": 0.7132667617689016,
814
+ "eval/nano_beir.nq": 0.39316532973343116,
815
+ "step": 1000
816
+ },
817
+ {
818
+ "epoch": 0.7132667617689016,
819
+ "eval/nano_beir.quora": 0.8295275476072758,
820
+ "step": 1000
821
+ },
822
+ {
823
+ "epoch": 0.7132667617689016,
824
+ "eval/nano_beir.scidocs": 0.2526832556333601,
825
+ "step": 1000
826
+ },
827
+ {
828
+ "epoch": 0.7132667617689016,
829
+ "eval/nano_beir.scifact": 0.7219633063518482,
830
+ "step": 1000
831
+ },
832
+ {
833
+ "epoch": 0.7132667617689016,
834
+ "eval/nano_beir.webis_touche2020": 0.2787587910246796,
835
+ "step": 1000
836
+ },
837
+ {
838
+ "epoch": 0.7132667617689016,
839
+ "eval/avg": 0.4571374679848476,
840
+ "step": 1000
841
+ },
842
+ {
843
+ "epoch": 0.7203994293865906,
844
+ "grad_norm": 5.15625,
845
+ "learning_rate": 9.68762555129222e-05,
846
+ "loss": 0.8872,
847
+ "step": 1010
848
+ },
849
+ {
850
+ "epoch": 0.7275320970042796,
851
+ "grad_norm": 5.65625,
852
+ "learning_rate": 9.675366799960841e-05,
853
+ "loss": 0.8334,
854
+ "step": 1020
855
+ },
856
+ {
857
+ "epoch": 0.7346647646219686,
858
+ "grad_norm": 5.1875,
859
+ "learning_rate": 9.662880177847594e-05,
860
+ "loss": 0.8014,
861
+ "step": 1030
862
+ },
863
+ {
864
+ "epoch": 0.7417974322396577,
865
+ "grad_norm": 4.9375,
866
+ "learning_rate": 9.65016629353282e-05,
867
+ "loss": 0.848,
868
+ "step": 1040
869
+ },
870
+ {
871
+ "epoch": 0.7489300998573466,
872
+ "grad_norm": 4.40625,
873
+ "learning_rate": 9.637225766673307e-05,
874
+ "loss": 0.7203,
875
+ "step": 1050
876
+ },
877
+ {
878
+ "epoch": 0.7560627674750356,
879
+ "grad_norm": 5.0625,
880
+ "learning_rate": 9.624059227972075e-05,
881
+ "loss": 0.8405,
882
+ "step": 1060
883
+ },
884
+ {
885
+ "epoch": 0.7631954350927247,
886
+ "grad_norm": 4.5625,
887
+ "learning_rate": 9.610667319147647e-05,
888
+ "loss": 0.8418,
889
+ "step": 1070
890
+ },
891
+ {
892
+ "epoch": 0.7703281027104137,
893
+ "grad_norm": 5.75,
894
+ "learning_rate": 9.597050692902765e-05,
895
+ "loss": 0.8853,
896
+ "step": 1080
897
+ },
898
+ {
899
+ "epoch": 0.7774607703281027,
900
+ "grad_norm": 4.09375,
901
+ "learning_rate": 9.583210012892582e-05,
902
+ "loss": 0.7733,
903
+ "step": 1090
904
+ },
905
+ {
906
+ "epoch": 0.7845934379457917,
907
+ "grad_norm": 4.40625,
908
+ "learning_rate": 9.569145953692315e-05,
909
+ "loss": 0.7698,
910
+ "step": 1100
911
+ },
912
+ {
913
+ "epoch": 0.7917261055634808,
914
+ "grad_norm": 4.84375,
915
+ "learning_rate": 9.55485920076437e-05,
916
+ "loss": 0.7759,
917
+ "step": 1110
918
+ },
919
+ {
920
+ "epoch": 0.7988587731811697,
921
+ "grad_norm": 5.90625,
922
+ "learning_rate": 9.540350450424926e-05,
923
+ "loss": 0.8025,
924
+ "step": 1120
925
+ },
926
+ {
927
+ "epoch": 0.8059914407988588,
928
+ "grad_norm": 5.53125,
929
+ "learning_rate": 9.525620409810007e-05,
930
+ "loss": 0.8364,
931
+ "step": 1130
932
+ },
933
+ {
934
+ "epoch": 0.8131241084165478,
935
+ "grad_norm": 5.03125,
936
+ "learning_rate": 9.510669796841014e-05,
937
+ "loss": 0.7523,
938
+ "step": 1140
939
+ },
940
+ {
941
+ "epoch": 0.8202567760342369,
942
+ "grad_norm": 5.46875,
943
+ "learning_rate": 9.495499340189728e-05,
944
+ "loss": 0.7333,
945
+ "step": 1150
946
+ },
947
+ {
948
+ "epoch": 0.8273894436519258,
949
+ "grad_norm": 4.875,
950
+ "learning_rate": 9.480109779242805e-05,
951
+ "loss": 0.7325,
952
+ "step": 1160
953
+ },
954
+ {
955
+ "epoch": 0.8345221112696148,
956
+ "grad_norm": 5.0,
957
+ "learning_rate": 9.464501864065735e-05,
958
+ "loss": 0.7996,
959
+ "step": 1170
960
+ },
961
+ {
962
+ "epoch": 0.8416547788873039,
963
+ "grad_norm": 4.6875,
964
+ "learning_rate": 9.448676355366282e-05,
965
+ "loss": 0.7946,
966
+ "step": 1180
967
+ },
968
+ {
969
+ "epoch": 0.8487874465049928,
970
+ "grad_norm": 5.21875,
971
+ "learning_rate": 9.432634024457414e-05,
972
+ "loss": 0.9133,
973
+ "step": 1190
974
+ },
975
+ {
976
+ "epoch": 0.8559201141226819,
977
+ "grad_norm": 4.78125,
978
+ "learning_rate": 9.416375653219709e-05,
979
+ "loss": 0.7218,
980
+ "step": 1200
981
+ },
982
+ {
983
+ "epoch": 0.8630527817403709,
984
+ "grad_norm": 4.28125,
985
+ "learning_rate": 9.399902034063242e-05,
986
+ "loss": 0.8453,
987
+ "step": 1210
988
+ },
989
+ {
990
+ "epoch": 0.8701854493580599,
991
+ "grad_norm": 5.1875,
992
+ "learning_rate": 9.383213969888972e-05,
993
+ "loss": 0.7261,
994
+ "step": 1220
995
+ },
996
+ {
997
+ "epoch": 0.8773181169757489,
998
+ "grad_norm": 3.828125,
999
+ "learning_rate": 9.366312274049602e-05,
1000
+ "loss": 0.6459,
1001
+ "step": 1230
1002
+ },
1003
+ {
1004
+ "epoch": 0.884450784593438,
1005
+ "grad_norm": 5.3125,
1006
+ "learning_rate": 9.34919777030994e-05,
1007
+ "loss": 0.7202,
1008
+ "step": 1240
1009
+ },
1010
+ {
1011
+ "epoch": 0.891583452211127,
1012
+ "grad_norm": 5.90625,
1013
+ "learning_rate": 9.331871292806759e-05,
1014
+ "loss": 0.7874,
1015
+ "step": 1250
1016
+ },
1017
+ {
1018
+ "epoch": 0.8987161198288159,
1019
+ "grad_norm": 4.09375,
1020
+ "learning_rate": 9.314333686008125e-05,
1021
+ "loss": 0.8524,
1022
+ "step": 1260
1023
+ },
1024
+ {
1025
+ "epoch": 0.905848787446505,
1026
+ "grad_norm": 5.75,
1027
+ "learning_rate": 9.296585804672252e-05,
1028
+ "loss": 0.7854,
1029
+ "step": 1270
1030
+ },
1031
+ {
1032
+ "epoch": 0.912981455064194,
1033
+ "grad_norm": 5.3125,
1034
+ "learning_rate": 9.278628513805838e-05,
1035
+ "loss": 0.8386,
1036
+ "step": 1280
1037
+ },
1038
+ {
1039
+ "epoch": 0.920114122681883,
1040
+ "grad_norm": 4.71875,
1041
+ "learning_rate": 9.260462688621905e-05,
1042
+ "loss": 0.8279,
1043
+ "step": 1290
1044
+ },
1045
+ {
1046
+ "epoch": 0.927246790299572,
1047
+ "grad_norm": 4.71875,
1048
+ "learning_rate": 9.242089214497145e-05,
1049
+ "loss": 0.8345,
1050
+ "step": 1300
1051
+ },
1052
+ {
1053
+ "epoch": 0.9343794579172611,
1054
+ "grad_norm": 5.15625,
1055
+ "learning_rate": 9.223508986928766e-05,
1056
+ "loss": 0.7596,
1057
+ "step": 1310
1058
+ },
1059
+ {
1060
+ "epoch": 0.9415121255349501,
1061
+ "grad_norm": 4.4375,
1062
+ "learning_rate": 9.204722911490846e-05,
1063
+ "loss": 0.6704,
1064
+ "step": 1320
1065
+ },
1066
+ {
1067
+ "epoch": 0.948644793152639,
1068
+ "grad_norm": 5.6875,
1069
+ "learning_rate": 9.1857319037902e-05,
1070
+ "loss": 0.7792,
1071
+ "step": 1330
1072
+ },
1073
+ {
1074
+ "epoch": 0.9557774607703281,
1075
+ "grad_norm": 4.78125,
1076
+ "learning_rate": 9.16653688942175e-05,
1077
+ "loss": 0.813,
1078
+ "step": 1340
1079
+ },
1080
+ {
1081
+ "epoch": 0.9629101283880172,
1082
+ "grad_norm": 4.4375,
1083
+ "learning_rate": 9.147138803923416e-05,
1084
+ "loss": 0.834,
1085
+ "step": 1350
1086
+ },
1087
+ {
1088
+ "epoch": 0.9700427960057061,
1089
+ "grad_norm": 4.1875,
1090
+ "learning_rate": 9.127538592730519e-05,
1091
+ "loss": 0.8299,
1092
+ "step": 1360
1093
+ },
1094
+ {
1095
+ "epoch": 0.9771754636233951,
1096
+ "grad_norm": 4.3125,
1097
+ "learning_rate": 9.107737211129701e-05,
1098
+ "loss": 0.7407,
1099
+ "step": 1370
1100
+ },
1101
+ {
1102
+ "epoch": 0.9843081312410842,
1103
+ "grad_norm": 5.0,
1104
+ "learning_rate": 9.087735624212365e-05,
1105
+ "loss": 0.8821,
1106
+ "step": 1380
1107
+ },
1108
+ {
1109
+ "epoch": 0.9914407988587732,
1110
+ "grad_norm": 4.71875,
1111
+ "learning_rate": 9.06753480682764e-05,
1112
+ "loss": 0.8595,
1113
+ "step": 1390
1114
+ },
1115
+ {
1116
+ "epoch": 0.9985734664764622,
1117
+ "grad_norm": 4.5,
1118
+ "learning_rate": 9.047135743534868e-05,
1119
+ "loss": 0.869,
1120
+ "step": 1400
1121
+ },
1122
+ {
1123
+ "epoch": 1.005706134094151,
1124
+ "grad_norm": 4.28125,
1125
+ "learning_rate": 9.02653942855561e-05,
1126
+ "loss": 0.8397,
1127
+ "step": 1410
1128
+ },
1129
+ {
1130
+ "epoch": 1.0128388017118402,
1131
+ "grad_norm": 5.34375,
1132
+ "learning_rate": 9.005746865725207e-05,
1133
+ "loss": 0.8047,
1134
+ "step": 1420
1135
+ },
1136
+ {
1137
+ "epoch": 1.0199714693295292,
1138
+ "grad_norm": 4.28125,
1139
+ "learning_rate": 8.984759068443831e-05,
1140
+ "loss": 0.72,
1141
+ "step": 1430
1142
+ },
1143
+ {
1144
+ "epoch": 1.0271041369472182,
1145
+ "grad_norm": 5.59375,
1146
+ "learning_rate": 8.963577059627118e-05,
1147
+ "loss": 0.7724,
1148
+ "step": 1440
1149
+ },
1150
+ {
1151
+ "epoch": 1.0342368045649073,
1152
+ "grad_norm": 4.3125,
1153
+ "learning_rate": 8.942201871656291e-05,
1154
+ "loss": 0.8014,
1155
+ "step": 1450
1156
+ },
1157
+ {
1158
+ "epoch": 1.0413694721825963,
1159
+ "grad_norm": 6.03125,
1160
+ "learning_rate": 8.920634546327858e-05,
1161
+ "loss": 0.8532,
1162
+ "step": 1460
1163
+ },
1164
+ {
1165
+ "epoch": 1.0485021398002854,
1166
+ "grad_norm": 3.75,
1167
+ "learning_rate": 8.898876134802826e-05,
1168
+ "loss": 0.7285,
1169
+ "step": 1470
1170
+ },
1171
+ {
1172
+ "epoch": 1.0556348074179742,
1173
+ "grad_norm": 4.5625,
1174
+ "learning_rate": 8.87692769755548e-05,
1175
+ "loss": 0.8574,
1176
+ "step": 1480
1177
+ },
1178
+ {
1179
+ "epoch": 1.0627674750356633,
1180
+ "grad_norm": 6.21875,
1181
+ "learning_rate": 8.854790304321681e-05,
1182
+ "loss": 0.76,
1183
+ "step": 1490
1184
+ },
1185
+ {
1186
+ "epoch": 1.0699001426533523,
1187
+ "grad_norm": 4.75,
1188
+ "learning_rate": 8.832465034046749e-05,
1189
+ "loss": 0.8441,
1190
+ "step": 1500
1191
+ },
1192
+ {
1193
+ "epoch": 1.0770328102710414,
1194
+ "grad_norm": 4.5625,
1195
+ "learning_rate": 8.80995297483286e-05,
1196
+ "loss": 0.764,
1197
+ "step": 1510
1198
+ },
1199
+ {
1200
+ "epoch": 1.0841654778887304,
1201
+ "grad_norm": 4.59375,
1202
+ "learning_rate": 8.787255223886019e-05,
1203
+ "loss": 0.7432,
1204
+ "step": 1520
1205
+ },
1206
+ {
1207
+ "epoch": 1.0912981455064195,
1208
+ "grad_norm": 4.1875,
1209
+ "learning_rate": 8.764372887462586e-05,
1210
+ "loss": 0.7894,
1211
+ "step": 1530
1212
+ },
1213
+ {
1214
+ "epoch": 1.0984308131241085,
1215
+ "grad_norm": 4.5,
1216
+ "learning_rate": 8.741307080815357e-05,
1217
+ "loss": 0.761,
1218
+ "step": 1540
1219
+ },
1220
+ {
1221
+ "epoch": 1.1055634807417973,
1222
+ "grad_norm": 6.03125,
1223
+ "learning_rate": 8.718058928139205e-05,
1224
+ "loss": 0.9432,
1225
+ "step": 1550
1226
+ },
1227
+ {
1228
+ "epoch": 1.1126961483594864,
1229
+ "grad_norm": 4.1875,
1230
+ "learning_rate": 8.694629562516294e-05,
1231
+ "loss": 0.7863,
1232
+ "step": 1560
1233
+ },
1234
+ {
1235
+ "epoch": 1.1198288159771754,
1236
+ "grad_norm": 4.21875,
1237
+ "learning_rate": 8.671020125860851e-05,
1238
+ "loss": 0.7732,
1239
+ "step": 1570
1240
+ },
1241
+ {
1242
+ "epoch": 1.1269614835948645,
1243
+ "grad_norm": 4.65625,
1244
+ "learning_rate": 8.647231768863512e-05,
1245
+ "loss": 0.7651,
1246
+ "step": 1580
1247
+ },
1248
+ {
1249
+ "epoch": 1.1340941512125535,
1250
+ "grad_norm": 4.34375,
1251
+ "learning_rate": 8.623265650935234e-05,
1252
+ "loss": 0.8021,
1253
+ "step": 1590
1254
+ },
1255
+ {
1256
+ "epoch": 1.1412268188302426,
1257
+ "grad_norm": 5.0625,
1258
+ "learning_rate": 8.599122940150795e-05,
1259
+ "loss": 0.8561,
1260
+ "step": 1600
1261
+ },
1262
+ {
1263
+ "epoch": 1.1483594864479316,
1264
+ "grad_norm": 5.03125,
1265
+ "learning_rate": 8.574804813191858e-05,
1266
+ "loss": 0.7551,
1267
+ "step": 1610
1268
+ },
1269
+ {
1270
+ "epoch": 1.1554921540656204,
1271
+ "grad_norm": 5.375,
1272
+ "learning_rate": 8.550312455289625e-05,
1273
+ "loss": 0.8045,
1274
+ "step": 1620
1275
+ },
1276
+ {
1277
+ "epoch": 1.1626248216833095,
1278
+ "grad_norm": 4.21875,
1279
+ "learning_rate": 8.525647060167063e-05,
1280
+ "loss": 0.7145,
1281
+ "step": 1630
1282
+ },
1283
+ {
1284
+ "epoch": 1.1697574893009985,
1285
+ "grad_norm": 4.125,
1286
+ "learning_rate": 8.500809829980734e-05,
1287
+ "loss": 0.8007,
1288
+ "step": 1640
1289
+ },
1290
+ {
1291
+ "epoch": 1.1768901569186876,
1292
+ "grad_norm": 5.03125,
1293
+ "learning_rate": 8.4758019752622e-05,
1294
+ "loss": 0.8009,
1295
+ "step": 1650
1296
+ },
1297
+ {
1298
+ "epoch": 1.1840228245363766,
1299
+ "grad_norm": 4.84375,
1300
+ "learning_rate": 8.450624714859016e-05,
1301
+ "loss": 0.833,
1302
+ "step": 1660
1303
+ },
1304
+ {
1305
+ "epoch": 1.1911554921540657,
1306
+ "grad_norm": 5.21875,
1307
+ "learning_rate": 8.425279275875336e-05,
1308
+ "loss": 0.8149,
1309
+ "step": 1670
1310
+ },
1311
+ {
1312
+ "epoch": 1.1982881597717547,
1313
+ "grad_norm": 4.1875,
1314
+ "learning_rate": 8.399766893612096e-05,
1315
+ "loss": 0.808,
1316
+ "step": 1680
1317
+ },
1318
+ {
1319
+ "epoch": 1.2054208273894436,
1320
+ "grad_norm": 4.625,
1321
+ "learning_rate": 8.374088811506819e-05,
1322
+ "loss": 0.7549,
1323
+ "step": 1690
1324
+ },
1325
+ {
1326
+ "epoch": 1.2125534950071326,
1327
+ "grad_norm": 5.71875,
1328
+ "learning_rate": 8.348246281072998e-05,
1329
+ "loss": 0.7755,
1330
+ "step": 1700
1331
+ },
1332
+ {
1333
+ "epoch": 1.2196861626248217,
1334
+ "grad_norm": 4.5625,
1335
+ "learning_rate": 8.322240561839109e-05,
1336
+ "loss": 0.8038,
1337
+ "step": 1710
1338
+ },
1339
+ {
1340
+ "epoch": 1.2268188302425107,
1341
+ "grad_norm": 4.09375,
1342
+ "learning_rate": 8.296072921287217e-05,
1343
+ "loss": 0.7207,
1344
+ "step": 1720
1345
+ },
1346
+ {
1347
+ "epoch": 1.2339514978601998,
1348
+ "grad_norm": 4.90625,
1349
+ "learning_rate": 8.269744634791208e-05,
1350
+ "loss": 0.7994,
1351
+ "step": 1730
1352
+ },
1353
+ {
1354
+ "epoch": 1.2410841654778888,
1355
+ "grad_norm": 4.3125,
1356
+ "learning_rate": 8.243256985554621e-05,
1357
+ "loss": 0.7989,
1358
+ "step": 1740
1359
+ },
1360
+ {
1361
+ "epoch": 1.2482168330955776,
1362
+ "grad_norm": 4.34375,
1363
+ "learning_rate": 8.21661126454811e-05,
1364
+ "loss": 0.6516,
1365
+ "step": 1750
1366
+ },
1367
+ {
1368
+ "epoch": 1.2553495007132667,
1369
+ "grad_norm": 4.59375,
1370
+ "learning_rate": 8.189808770446528e-05,
1371
+ "loss": 0.7919,
1372
+ "step": 1760
1373
+ },
1374
+ {
1375
+ "epoch": 1.2624821683309557,
1376
+ "grad_norm": 5.53125,
1377
+ "learning_rate": 8.162850809565623e-05,
1378
+ "loss": 0.9372,
1379
+ "step": 1770
1380
+ },
1381
+ {
1382
+ "epoch": 1.2696148359486448,
1383
+ "grad_norm": 4.3125,
1384
+ "learning_rate": 8.135738695798376e-05,
1385
+ "loss": 0.7683,
1386
+ "step": 1780
1387
+ },
1388
+ {
1389
+ "epoch": 1.2767475035663338,
1390
+ "grad_norm": 4.3125,
1391
+ "learning_rate": 8.108473750550965e-05,
1392
+ "loss": 0.737,
1393
+ "step": 1790
1394
+ },
1395
+ {
1396
+ "epoch": 1.2838801711840229,
1397
+ "grad_norm": 5.78125,
1398
+ "learning_rate": 8.081057302678352e-05,
1399
+ "loss": 0.7507,
1400
+ "step": 1800
1401
+ },
1402
+ {
1403
+ "epoch": 1.291012838801712,
1404
+ "grad_norm": 4.875,
1405
+ "learning_rate": 8.053490688419533e-05,
1406
+ "loss": 0.7016,
1407
+ "step": 1810
1408
+ },
1409
+ {
1410
+ "epoch": 1.298145506419401,
1411
+ "grad_norm": 4.5,
1412
+ "learning_rate": 8.02577525133239e-05,
1413
+ "loss": 0.7151,
1414
+ "step": 1820
1415
+ },
1416
+ {
1417
+ "epoch": 1.3052781740370898,
1418
+ "grad_norm": 4.5625,
1419
+ "learning_rate": 7.997912342228232e-05,
1420
+ "loss": 0.8559,
1421
+ "step": 1830
1422
+ },
1423
+ {
1424
+ "epoch": 1.3124108416547788,
1425
+ "grad_norm": 4.15625,
1426
+ "learning_rate": 7.969903319105935e-05,
1427
+ "loss": 0.7118,
1428
+ "step": 1840
1429
+ },
1430
+ {
1431
+ "epoch": 1.3195435092724679,
1432
+ "grad_norm": 3.78125,
1433
+ "learning_rate": 7.941749547085777e-05,
1434
+ "loss": 0.7855,
1435
+ "step": 1850
1436
+ },
1437
+ {
1438
+ "epoch": 1.326676176890157,
1439
+ "grad_norm": 4.53125,
1440
+ "learning_rate": 7.913452398342881e-05,
1441
+ "loss": 0.7438,
1442
+ "step": 1860
1443
+ },
1444
+ {
1445
+ "epoch": 1.333808844507846,
1446
+ "grad_norm": 4.65625,
1447
+ "learning_rate": 7.885013252040359e-05,
1448
+ "loss": 0.7018,
1449
+ "step": 1870
1450
+ },
1451
+ {
1452
+ "epoch": 1.340941512125535,
1453
+ "grad_norm": 6.9375,
1454
+ "learning_rate": 7.856433494262077e-05,
1455
+ "loss": 0.7083,
1456
+ "step": 1880
1457
+ },
1458
+ {
1459
+ "epoch": 1.3480741797432239,
1460
+ "grad_norm": 4.875,
1461
+ "learning_rate": 7.827714517945115e-05,
1462
+ "loss": 0.8265,
1463
+ "step": 1890
1464
+ },
1465
+ {
1466
+ "epoch": 1.355206847360913,
1467
+ "grad_norm": 5.5625,
1468
+ "learning_rate": 7.798857722811856e-05,
1469
+ "loss": 0.8735,
1470
+ "step": 1900
1471
+ },
1472
+ {
1473
+ "epoch": 1.362339514978602,
1474
+ "grad_norm": 5.71875,
1475
+ "learning_rate": 7.769864515301787e-05,
1476
+ "loss": 0.7256,
1477
+ "step": 1910
1478
+ },
1479
+ {
1480
+ "epoch": 1.369472182596291,
1481
+ "grad_norm": 4.53125,
1482
+ "learning_rate": 7.740736308502938e-05,
1483
+ "loss": 0.7472,
1484
+ "step": 1920
1485
+ },
1486
+ {
1487
+ "epoch": 1.37660485021398,
1488
+ "grad_norm": 5.875,
1489
+ "learning_rate": 7.711474522083015e-05,
1490
+ "loss": 0.7601,
1491
+ "step": 1930
1492
+ },
1493
+ {
1494
+ "epoch": 1.383737517831669,
1495
+ "grad_norm": 5.75,
1496
+ "learning_rate": 7.682080582220206e-05,
1497
+ "loss": 0.8857,
1498
+ "step": 1940
1499
+ },
1500
+ {
1501
+ "epoch": 1.3908701854493581,
1502
+ "grad_norm": 6.40625,
1503
+ "learning_rate": 7.65255592153367e-05,
1504
+ "loss": 0.7296,
1505
+ "step": 1950
1506
+ },
1507
+ {
1508
+ "epoch": 1.3980028530670472,
1509
+ "grad_norm": 4.40625,
1510
+ "learning_rate": 7.622901979013717e-05,
1511
+ "loss": 0.7331,
1512
+ "step": 1960
1513
+ },
1514
+ {
1515
+ "epoch": 1.405135520684736,
1516
+ "grad_norm": 5.09375,
1517
+ "learning_rate": 7.593120199951671e-05,
1518
+ "loss": 0.8462,
1519
+ "step": 1970
1520
+ },
1521
+ {
1522
+ "epoch": 1.412268188302425,
1523
+ "grad_norm": 4.15625,
1524
+ "learning_rate": 7.563212035869425e-05,
1525
+ "loss": 0.7307,
1526
+ "step": 1980
1527
+ },
1528
+ {
1529
+ "epoch": 1.4194008559201141,
1530
+ "grad_norm": 4.90625,
1531
+ "learning_rate": 7.533178944448705e-05,
1532
+ "loss": 0.7711,
1533
+ "step": 1990
1534
+ },
1535
+ {
1536
+ "epoch": 1.4265335235378032,
1537
+ "grad_norm": 5.78125,
1538
+ "learning_rate": 7.503022389460014e-05,
1539
+ "loss": 0.7485,
1540
+ "step": 2000
1541
+ },
1542
+ {
1543
+ "epoch": 1.4265335235378032,
1544
+ "eval/nano_beir.arguana": 0.47494654864753366,
1545
+ "step": 2000
1546
+ },
1547
+ {
1548
+ "epoch": 1.4265335235378032,
1549
+ "eval/nano_beir.climate_fever": 0.2496059194905162,
1550
+ "step": 2000
1551
+ },
1552
+ {
1553
+ "epoch": 1.4265335235378032,
1554
+ "eval/nano_beir.dbpedia_entity": 0.3371368383795953,
1555
+ "step": 2000
1556
+ },
1557
+ {
1558
+ "epoch": 1.4265335235378032,
1559
+ "eval/nano_beir.fever": 0.6065422117591313,
1560
+ "step": 2000
1561
+ },
1562
+ {
1563
+ "epoch": 1.4265335235378032,
1564
+ "eval/nano_beir.fiqa": 0.4135337243548038,
1565
+ "step": 2000
1566
+ },
1567
+ {
1568
+ "epoch": 1.4265335235378032,
1569
+ "eval/nano_beir.hotpotqa": 0.6804056480114672,
1570
+ "step": 2000
1571
+ },
1572
+ {
1573
+ "epoch": 1.4265335235378032,
1574
+ "eval/nano_beir.nfcorpus": 0.35238501477624823,
1575
+ "step": 2000
1576
+ },
1577
+ {
1578
+ "epoch": 1.4265335235378032,
1579
+ "eval/nano_beir.nq": 0.39770361480597743,
1580
+ "step": 2000
1581
+ },
1582
+ {
1583
+ "epoch": 1.4265335235378032,
1584
+ "eval/nano_beir.quora": 0.8217479889085385,
1585
+ "step": 2000
1586
+ },
1587
+ {
1588
+ "epoch": 1.4265335235378032,
1589
+ "eval/nano_beir.scidocs": 0.2592573629546752,
1590
+ "step": 2000
1591
+ },
1592
+ {
1593
+ "epoch": 1.4265335235378032,
1594
+ "eval/nano_beir.scifact": 0.7479618950231317,
1595
+ "step": 2000
1596
+ },
1597
+ {
1598
+ "epoch": 1.4265335235378032,
1599
+ "eval/nano_beir.webis_touche2020": 0.2719065772753103,
1600
+ "step": 2000
1601
+ },
1602
+ {
1603
+ "epoch": 1.4265335235378032,
1604
+ "eval/avg": 0.467761112032244,
1605
+ "step": 2000
1606
+ },
1607
+ {
1608
+ "epoch": 1.4336661911554922,
1609
+ "grad_norm": 3.71875,
1610
+ "learning_rate": 7.472743840691299e-05,
1611
+ "loss": 0.8505,
1612
+ "step": 2010
1613
+ },
1614
+ {
1615
+ "epoch": 1.440798858773181,
1616
+ "grad_norm": 6.15625,
1617
+ "learning_rate": 7.44234477387631e-05,
1618
+ "loss": 0.7449,
1619
+ "step": 2020
1620
+ },
1621
+ {
1622
+ "epoch": 1.44793152639087,
1623
+ "grad_norm": 4.0625,
1624
+ "learning_rate": 7.411826670622676e-05,
1625
+ "loss": 0.6961,
1626
+ "step": 2030
1627
+ },
1628
+ {
1629
+ "epoch": 1.4550641940085591,
1630
+ "grad_norm": 4.40625,
1631
+ "learning_rate": 7.381191018339696e-05,
1632
+ "loss": 0.8105,
1633
+ "step": 2040
1634
+ },
1635
+ {
1636
+ "epoch": 1.4621968616262482,
1637
+ "grad_norm": 5.59375,
1638
+ "learning_rate": 7.350439310165842e-05,
1639
+ "loss": 0.8111,
1640
+ "step": 2050
1641
+ },
1642
+ {
1643
+ "epoch": 1.4693295292439372,
1644
+ "grad_norm": 4.625,
1645
+ "learning_rate": 7.319573044895986e-05,
1646
+ "loss": 0.7972,
1647
+ "step": 2060
1648
+ },
1649
+ {
1650
+ "epoch": 1.4764621968616263,
1651
+ "grad_norm": 3.953125,
1652
+ "learning_rate": 7.28859372690835e-05,
1653
+ "loss": 0.7745,
1654
+ "step": 2070
1655
+ },
1656
+ {
1657
+ "epoch": 1.4835948644793153,
1658
+ "grad_norm": 3.859375,
1659
+ "learning_rate": 7.257502866091192e-05,
1660
+ "loss": 0.7904,
1661
+ "step": 2080
1662
+ },
1663
+ {
1664
+ "epoch": 1.4907275320970044,
1665
+ "grad_norm": 4.875,
1666
+ "learning_rate": 7.226301977769199e-05,
1667
+ "loss": 0.7724,
1668
+ "step": 2090
1669
+ },
1670
+ {
1671
+ "epoch": 1.4978601997146934,
1672
+ "grad_norm": 5.59375,
1673
+ "learning_rate": 7.194992582629654e-05,
1674
+ "loss": 0.7693,
1675
+ "step": 2100
1676
+ },
1677
+ {
1678
+ "epoch": 1.5049928673323825,
1679
+ "grad_norm": 4.75,
1680
+ "learning_rate": 7.163576206648304e-05,
1681
+ "loss": 0.6929,
1682
+ "step": 2110
1683
+ },
1684
+ {
1685
+ "epoch": 1.5121255349500713,
1686
+ "grad_norm": 5.78125,
1687
+ "learning_rate": 7.132054381014995e-05,
1688
+ "loss": 0.8008,
1689
+ "step": 2120
1690
+ },
1691
+ {
1692
+ "epoch": 1.5192582025677603,
1693
+ "grad_norm": 6.28125,
1694
+ "learning_rate": 7.100428642059033e-05,
1695
+ "loss": 0.8399,
1696
+ "step": 2130
1697
+ },
1698
+ {
1699
+ "epoch": 1.5263908701854494,
1700
+ "grad_norm": 6.3125,
1701
+ "learning_rate": 7.068700531174319e-05,
1702
+ "loss": 0.7399,
1703
+ "step": 2140
1704
+ },
1705
+ {
1706
+ "epoch": 1.5335235378031382,
1707
+ "grad_norm": 4.1875,
1708
+ "learning_rate": 7.036871594744217e-05,
1709
+ "loss": 0.7796,
1710
+ "step": 2150
1711
+ },
1712
+ {
1713
+ "epoch": 1.5406562054208273,
1714
+ "grad_norm": 5.34375,
1715
+ "learning_rate": 7.004943384066187e-05,
1716
+ "loss": 0.8685,
1717
+ "step": 2160
1718
+ },
1719
+ {
1720
+ "epoch": 1.5477888730385163,
1721
+ "grad_norm": 4.8125,
1722
+ "learning_rate": 6.97291745527617e-05,
1723
+ "loss": 0.7448,
1724
+ "step": 2170
1725
+ },
1726
+ {
1727
+ "epoch": 1.5549215406562054,
1728
+ "grad_norm": 6.09375,
1729
+ "learning_rate": 6.940795369272754e-05,
1730
+ "loss": 0.8527,
1731
+ "step": 2180
1732
+ },
1733
+ {
1734
+ "epoch": 1.5620542082738944,
1735
+ "grad_norm": 5.15625,
1736
+ "learning_rate": 6.908578691641092e-05,
1737
+ "loss": 0.8704,
1738
+ "step": 2190
1739
+ },
1740
+ {
1741
+ "epoch": 1.5691868758915835,
1742
+ "grad_norm": 4.3125,
1743
+ "learning_rate": 6.876268992576604e-05,
1744
+ "loss": 0.7033,
1745
+ "step": 2200
1746
+ },
1747
+ {
1748
+ "epoch": 1.5763195435092725,
1749
+ "grad_norm": 5.6875,
1750
+ "learning_rate": 6.843867846808438e-05,
1751
+ "loss": 0.7664,
1752
+ "step": 2210
1753
+ },
1754
+ {
1755
+ "epoch": 1.5834522111269616,
1756
+ "grad_norm": 4.5625,
1757
+ "learning_rate": 6.811376833522729e-05,
1758
+ "loss": 0.7902,
1759
+ "step": 2220
1760
+ },
1761
+ {
1762
+ "epoch": 1.5905848787446506,
1763
+ "grad_norm": 4.6875,
1764
+ "learning_rate": 6.778797536285624e-05,
1765
+ "loss": 0.7585,
1766
+ "step": 2230
1767
+ },
1768
+ {
1769
+ "epoch": 1.5977175463623396,
1770
+ "grad_norm": 5.375,
1771
+ "learning_rate": 6.746131542966112e-05,
1772
+ "loss": 0.8124,
1773
+ "step": 2240
1774
+ },
1775
+ {
1776
+ "epoch": 1.6048502139800287,
1777
+ "grad_norm": 4.59375,
1778
+ "learning_rate": 6.713380445658618e-05,
1779
+ "loss": 0.7936,
1780
+ "step": 2250
1781
+ },
1782
+ {
1783
+ "epoch": 1.6119828815977175,
1784
+ "grad_norm": 3.5625,
1785
+ "learning_rate": 6.680545840605423e-05,
1786
+ "loss": 0.6785,
1787
+ "step": 2260
1788
+ },
1789
+ {
1790
+ "epoch": 1.6191155492154066,
1791
+ "grad_norm": 5.875,
1792
+ "learning_rate": 6.647629328118851e-05,
1793
+ "loss": 0.8321,
1794
+ "step": 2270
1795
+ },
1796
+ {
1797
+ "epoch": 1.6262482168330956,
1798
+ "grad_norm": 5.125,
1799
+ "learning_rate": 6.614632512503288e-05,
1800
+ "loss": 0.8184,
1801
+ "step": 2280
1802
+ },
1803
+ {
1804
+ "epoch": 1.6333808844507844,
1805
+ "grad_norm": 4.5,
1806
+ "learning_rate": 6.58155700197697e-05,
1807
+ "loss": 0.7187,
1808
+ "step": 2290
1809
+ },
1810
+ {
1811
+ "epoch": 1.6405135520684735,
1812
+ "grad_norm": 4.96875,
1813
+ "learning_rate": 6.548404408593621e-05,
1814
+ "loss": 0.6873,
1815
+ "step": 2300
1816
+ },
1817
+ {
1818
+ "epoch": 1.6476462196861625,
1819
+ "grad_norm": 4.96875,
1820
+ "learning_rate": 6.515176348163871e-05,
1821
+ "loss": 0.7537,
1822
+ "step": 2310
1823
+ },
1824
+ {
1825
+ "epoch": 1.6547788873038516,
1826
+ "grad_norm": 4.71875,
1827
+ "learning_rate": 6.481874440176506e-05,
1828
+ "loss": 0.8058,
1829
+ "step": 2320
1830
+ },
1831
+ {
1832
+ "epoch": 1.6619115549215406,
1833
+ "grad_norm": 5.375,
1834
+ "learning_rate": 6.448500307719538e-05,
1835
+ "loss": 0.8117,
1836
+ "step": 2330
1837
+ },
1838
+ {
1839
+ "epoch": 1.6690442225392297,
1840
+ "grad_norm": 5.15625,
1841
+ "learning_rate": 6.415055577401102e-05,
1842
+ "loss": 0.7938,
1843
+ "step": 2340
1844
+ },
1845
+ {
1846
+ "epoch": 1.6761768901569187,
1847
+ "grad_norm": 6.15625,
1848
+ "learning_rate": 6.381541879270169e-05,
1849
+ "loss": 0.7898,
1850
+ "step": 2350
1851
+ },
1852
+ {
1853
+ "epoch": 1.6833095577746078,
1854
+ "grad_norm": 4.03125,
1855
+ "learning_rate": 6.347960846737106e-05,
1856
+ "loss": 0.7896,
1857
+ "step": 2360
1858
+ },
1859
+ {
1860
+ "epoch": 1.6904422253922968,
1861
+ "grad_norm": 4.28125,
1862
+ "learning_rate": 6.31431411649406e-05,
1863
+ "loss": 0.6994,
1864
+ "step": 2370
1865
+ },
1866
+ {
1867
+ "epoch": 1.6975748930099859,
1868
+ "grad_norm": 5.71875,
1869
+ "learning_rate": 6.280603328435198e-05,
1870
+ "loss": 0.7416,
1871
+ "step": 2380
1872
+ },
1873
+ {
1874
+ "epoch": 1.7047075606276747,
1875
+ "grad_norm": 4.8125,
1876
+ "learning_rate": 6.24683012557677e-05,
1877
+ "loss": 0.7143,
1878
+ "step": 2390
1879
+ },
1880
+ {
1881
+ "epoch": 1.7118402282453637,
1882
+ "grad_norm": 5.28125,
1883
+ "learning_rate": 6.212996153977037e-05,
1884
+ "loss": 0.6699,
1885
+ "step": 2400
1886
+ },
1887
+ {
1888
+ "epoch": 1.7189728958630528,
1889
+ "grad_norm": 4.84375,
1890
+ "learning_rate": 6.179103062656042e-05,
1891
+ "loss": 0.7084,
1892
+ "step": 2410
1893
+ },
1894
+ {
1895
+ "epoch": 1.7261055634807418,
1896
+ "grad_norm": 5.125,
1897
+ "learning_rate": 6.145152503515239e-05,
1898
+ "loss": 0.7638,
1899
+ "step": 2420
1900
+ },
1901
+ {
1902
+ "epoch": 1.7332382310984307,
1903
+ "grad_norm": 5.0,
1904
+ "learning_rate": 6.111146131256983e-05,
1905
+ "loss": 0.7875,
1906
+ "step": 2430
1907
+ },
1908
+ {
1909
+ "epoch": 1.7403708987161197,
1910
+ "grad_norm": 3.921875,
1911
+ "learning_rate": 6.077085603303882e-05,
1912
+ "loss": 0.8864,
1913
+ "step": 2440
1914
+ },
1915
+ {
1916
+ "epoch": 1.7475035663338088,
1917
+ "grad_norm": 4.84375,
1918
+ "learning_rate": 6.042972579718019e-05,
1919
+ "loss": 0.7354,
1920
+ "step": 2450
1921
+ },
1922
+ {
1923
+ "epoch": 1.7546362339514978,
1924
+ "grad_norm": 3.90625,
1925
+ "learning_rate": 6.008808723120035e-05,
1926
+ "loss": 0.7581,
1927
+ "step": 2460
1928
+ },
1929
+ {
1930
+ "epoch": 1.7617689015691869,
1931
+ "grad_norm": 4.40625,
1932
+ "learning_rate": 5.974595698608103e-05,
1933
+ "loss": 0.738,
1934
+ "step": 2470
1935
+ },
1936
+ {
1937
+ "epoch": 1.768901569186876,
1938
+ "grad_norm": 4.53125,
1939
+ "learning_rate": 5.9403351736767696e-05,
1940
+ "loss": 0.6686,
1941
+ "step": 2480
1942
+ },
1943
+ {
1944
+ "epoch": 1.776034236804565,
1945
+ "grad_norm": 5.125,
1946
+ "learning_rate": 5.906028818135687e-05,
1947
+ "loss": 0.6354,
1948
+ "step": 2490
1949
+ },
1950
+ {
1951
+ "epoch": 1.783166904422254,
1952
+ "grad_norm": 5.125,
1953
+ "learning_rate": 5.8716783040282244e-05,
1954
+ "loss": 0.7869,
1955
+ "step": 2500
1956
+ },
1957
+ {
1958
+ "epoch": 1.790299572039943,
1959
+ "grad_norm": 4.21875,
1960
+ "learning_rate": 5.837285305549978e-05,
1961
+ "loss": 0.7463,
1962
+ "step": 2510
1963
+ },
1964
+ {
1965
+ "epoch": 1.797432239657632,
1966
+ "grad_norm": 3.9375,
1967
+ "learning_rate": 5.8028514989671724e-05,
1968
+ "loss": 0.7524,
1969
+ "step": 2520
1970
+ },
1971
+ {
1972
+ "epoch": 1.804564907275321,
1973
+ "grad_norm": 6.15625,
1974
+ "learning_rate": 5.7683785625349616e-05,
1975
+ "loss": 0.7558,
1976
+ "step": 2530
1977
+ },
1978
+ {
1979
+ "epoch": 1.81169757489301,
1980
+ "grad_norm": 4.40625,
1981
+ "learning_rate": 5.733868176415633e-05,
1982
+ "loss": 0.8002,
1983
+ "step": 2540
1984
+ },
1985
+ {
1986
+ "epoch": 1.818830242510699,
1987
+ "grad_norm": 5.5625,
1988
+ "learning_rate": 5.699322022596722e-05,
1989
+ "loss": 0.7829,
1990
+ "step": 2550
1991
+ },
1992
+ {
1993
+ "epoch": 1.825962910128388,
1994
+ "grad_norm": 4.375,
1995
+ "learning_rate": 5.664741784809022e-05,
1996
+ "loss": 0.7097,
1997
+ "step": 2560
1998
+ },
1999
+ {
2000
+ "epoch": 1.833095577746077,
2001
+ "grad_norm": 4.96875,
2002
+ "learning_rate": 5.630129148444543e-05,
2003
+ "loss": 0.6793,
2004
+ "step": 2570
2005
+ },
2006
+ {
2007
+ "epoch": 1.840228245363766,
2008
+ "grad_norm": 4.84375,
2009
+ "learning_rate": 5.595485800474349e-05,
2010
+ "loss": 0.7081,
2011
+ "step": 2580
2012
+ },
2013
+ {
2014
+ "epoch": 1.847360912981455,
2015
+ "grad_norm": 4.25,
2016
+ "learning_rate": 5.560813429366345e-05,
2017
+ "loss": 0.911,
2018
+ "step": 2590
2019
+ },
2020
+ {
2021
+ "epoch": 1.854493580599144,
2022
+ "grad_norm": 5.5625,
2023
+ "learning_rate": 5.5261137250029835e-05,
2024
+ "loss": 0.8128,
2025
+ "step": 2600
2026
+ },
2027
+ {
2028
+ "epoch": 1.861626248216833,
2029
+ "grad_norm": 4.34375,
2030
+ "learning_rate": 5.4913883785988993e-05,
2031
+ "loss": 0.7784,
2032
+ "step": 2610
2033
+ },
2034
+ {
2035
+ "epoch": 1.8687589158345221,
2036
+ "grad_norm": 3.859375,
2037
+ "learning_rate": 5.456639082618489e-05,
2038
+ "loss": 0.6765,
2039
+ "step": 2620
2040
+ },
2041
+ {
2042
+ "epoch": 1.8758915834522112,
2043
+ "grad_norm": 4.96875,
2044
+ "learning_rate": 5.421867530693414e-05,
2045
+ "loss": 0.7464,
2046
+ "step": 2630
2047
+ },
2048
+ {
2049
+ "epoch": 1.8830242510699002,
2050
+ "grad_norm": 4.4375,
2051
+ "learning_rate": 5.3870754175400595e-05,
2052
+ "loss": 0.7452,
2053
+ "step": 2640
2054
+ },
2055
+ {
2056
+ "epoch": 1.8901569186875893,
2057
+ "grad_norm": 4.40625,
2058
+ "learning_rate": 5.352264438876935e-05,
2059
+ "loss": 0.7475,
2060
+ "step": 2650
2061
+ },
2062
+ {
2063
+ "epoch": 1.8972895863052783,
2064
+ "grad_norm": 3.6875,
2065
+ "learning_rate": 5.3174362913420306e-05,
2066
+ "loss": 0.761,
2067
+ "step": 2660
2068
+ },
2069
+ {
2070
+ "epoch": 1.9044222539229672,
2071
+ "grad_norm": 4.5625,
2072
+ "learning_rate": 5.2825926724101236e-05,
2073
+ "loss": 0.7183,
2074
+ "step": 2670
2075
+ },
2076
+ {
2077
+ "epoch": 1.9115549215406562,
2078
+ "grad_norm": 4.875,
2079
+ "learning_rate": 5.247735280310041e-05,
2080
+ "loss": 0.756,
2081
+ "step": 2680
2082
+ },
2083
+ {
2084
+ "epoch": 1.9186875891583453,
2085
+ "grad_norm": 4.8125,
2086
+ "learning_rate": 5.212865813941899e-05,
2087
+ "loss": 0.6997,
2088
+ "step": 2690
2089
+ },
2090
+ {
2091
+ "epoch": 1.925820256776034,
2092
+ "grad_norm": 4.75,
2093
+ "learning_rate": 5.1779859727942924e-05,
2094
+ "loss": 0.7978,
2095
+ "step": 2700
2096
+ },
2097
+ {
2098
+ "epoch": 1.9329529243937231,
2099
+ "grad_norm": 4.5,
2100
+ "learning_rate": 5.143097456861474e-05,
2101
+ "loss": 0.7099,
2102
+ "step": 2710
2103
+ },
2104
+ {
2105
+ "epoch": 1.9400855920114122,
2106
+ "grad_norm": 4.8125,
2107
+ "learning_rate": 5.1082019665604895e-05,
2108
+ "loss": 0.7029,
2109
+ "step": 2720
2110
+ },
2111
+ {
2112
+ "epoch": 1.9472182596291012,
2113
+ "grad_norm": 5.875,
2114
+ "learning_rate": 5.073301202648304e-05,
2115
+ "loss": 0.7328,
2116
+ "step": 2730
2117
+ },
2118
+ {
2119
+ "epoch": 1.9543509272467903,
2120
+ "grad_norm": 3.875,
2121
+ "learning_rate": 5.0383968661389146e-05,
2122
+ "loss": 0.7982,
2123
+ "step": 2740
2124
+ },
2125
+ {
2126
+ "epoch": 1.9614835948644793,
2127
+ "grad_norm": 4.71875,
2128
+ "learning_rate": 5.003490658220438e-05,
2129
+ "loss": 0.7126,
2130
+ "step": 2750
2131
+ },
2132
+ {
2133
+ "epoch": 1.9686162624821684,
2134
+ "grad_norm": 4.125,
2135
+ "learning_rate": 4.968584280172206e-05,
2136
+ "loss": 0.6598,
2137
+ "step": 2760
2138
+ },
2139
+ {
2140
+ "epoch": 1.9757489300998574,
2141
+ "grad_norm": 4.03125,
2142
+ "learning_rate": 4.933679433281836e-05,
2143
+ "loss": 0.7039,
2144
+ "step": 2770
2145
+ },
2146
+ {
2147
+ "epoch": 1.9828815977175465,
2148
+ "grad_norm": 4.78125,
2149
+ "learning_rate": 4.8987778187623245e-05,
2150
+ "loss": 0.6858,
2151
+ "step": 2780
2152
+ },
2153
+ {
2154
+ "epoch": 1.9900142653352355,
2155
+ "grad_norm": 4.53125,
2156
+ "learning_rate": 4.863881137669123e-05,
2157
+ "loss": 0.7924,
2158
+ "step": 2790
2159
+ },
2160
+ {
2161
+ "epoch": 1.9971469329529246,
2162
+ "grad_norm": 5.40625,
2163
+ "learning_rate": 4.8289910908172376e-05,
2164
+ "loss": 0.7606,
2165
+ "step": 2800
2166
+ },
2167
+ {
2168
+ "epoch": 2.0042796005706136,
2169
+ "grad_norm": 5.0,
2170
+ "learning_rate": 4.794109378698327e-05,
2171
+ "loss": 0.7584,
2172
+ "step": 2810
2173
+ },
2174
+ {
2175
+ "epoch": 2.011412268188302,
2176
+ "grad_norm": 5.34375,
2177
+ "learning_rate": 4.7592377013978306e-05,
2178
+ "loss": 0.6966,
2179
+ "step": 2820
2180
+ },
2181
+ {
2182
+ "epoch": 2.0185449358059913,
2183
+ "grad_norm": 4.5625,
2184
+ "learning_rate": 4.724377758512103e-05,
2185
+ "loss": 0.8236,
2186
+ "step": 2830
2187
+ },
2188
+ {
2189
+ "epoch": 2.0256776034236803,
2190
+ "grad_norm": 4.1875,
2191
+ "learning_rate": 4.68953124906558e-05,
2192
+ "loss": 0.7729,
2193
+ "step": 2840
2194
+ },
2195
+ {
2196
+ "epoch": 2.0328102710413694,
2197
+ "grad_norm": 4.875,
2198
+ "learning_rate": 4.654699871427971e-05,
2199
+ "loss": 0.6784,
2200
+ "step": 2850
2201
+ },
2202
+ {
2203
+ "epoch": 2.0399429386590584,
2204
+ "grad_norm": 8.8125,
2205
+ "learning_rate": 4.619885323231484e-05,
2206
+ "loss": 0.8262,
2207
+ "step": 2860
2208
+ },
2209
+ {
2210
+ "epoch": 2.0470756062767475,
2211
+ "grad_norm": 4.15625,
2212
+ "learning_rate": 4.58508930128808e-05,
2213
+ "loss": 0.7451,
2214
+ "step": 2870
2215
+ },
2216
+ {
2217
+ "epoch": 2.0542082738944365,
2218
+ "grad_norm": 5.59375,
2219
+ "learning_rate": 4.550313501506781e-05,
2220
+ "loss": 0.7622,
2221
+ "step": 2880
2222
+ },
2223
+ {
2224
+ "epoch": 2.0613409415121255,
2225
+ "grad_norm": 5.9375,
2226
+ "learning_rate": 4.515559618811005e-05,
2227
+ "loss": 0.6851,
2228
+ "step": 2890
2229
+ },
2230
+ {
2231
+ "epoch": 2.0684736091298146,
2232
+ "grad_norm": 4.375,
2233
+ "learning_rate": 4.4808293470559643e-05,
2234
+ "loss": 0.7706,
2235
+ "step": 2900
2236
+ },
2237
+ {
2238
+ "epoch": 2.0756062767475036,
2239
+ "grad_norm": 4.84375,
2240
+ "learning_rate": 4.446124378946107e-05,
2241
+ "loss": 0.7432,
2242
+ "step": 2910
2243
+ },
2244
+ {
2245
+ "epoch": 2.0827389443651927,
2246
+ "grad_norm": 4.21875,
2247
+ "learning_rate": 4.411446405952618e-05,
2248
+ "loss": 0.7158,
2249
+ "step": 2920
2250
+ },
2251
+ {
2252
+ "epoch": 2.0898716119828817,
2253
+ "grad_norm": 4.375,
2254
+ "learning_rate": 4.376797118230979e-05,
2255
+ "loss": 0.7035,
2256
+ "step": 2930
2257
+ },
2258
+ {
2259
+ "epoch": 2.097004279600571,
2260
+ "grad_norm": 4.65625,
2261
+ "learning_rate": 4.342178204538588e-05,
2262
+ "loss": 0.7611,
2263
+ "step": 2940
2264
+ },
2265
+ {
2266
+ "epoch": 2.10413694721826,
2267
+ "grad_norm": 4.625,
2268
+ "learning_rate": 4.307591352152459e-05,
2269
+ "loss": 0.6861,
2270
+ "step": 2950
2271
+ },
2272
+ {
2273
+ "epoch": 2.1112696148359484,
2274
+ "grad_norm": 5.875,
2275
+ "learning_rate": 4.273038246786986e-05,
2276
+ "loss": 0.8394,
2277
+ "step": 2960
2278
+ },
2279
+ {
2280
+ "epoch": 2.1184022824536375,
2281
+ "grad_norm": 4.1875,
2282
+ "learning_rate": 4.238520572511773e-05,
2283
+ "loss": 0.6647,
2284
+ "step": 2970
2285
+ },
2286
+ {
2287
+ "epoch": 2.1255349500713265,
2288
+ "grad_norm": 4.3125,
2289
+ "learning_rate": 4.2040400116695666e-05,
2290
+ "loss": 0.7305,
2291
+ "step": 2980
2292
+ },
2293
+ {
2294
+ "epoch": 2.1326676176890156,
2295
+ "grad_norm": 5.59375,
2296
+ "learning_rate": 4.169598244794261e-05,
2297
+ "loss": 0.7693,
2298
+ "step": 2990
2299
+ },
2300
+ {
2301
+ "epoch": 2.1398002853067046,
2302
+ "grad_norm": 5.28125,
2303
+ "learning_rate": 4.135196950528982e-05,
2304
+ "loss": 0.7392,
2305
+ "step": 3000
2306
+ },
2307
+ {
2308
+ "epoch": 2.1398002853067046,
2309
+ "eval/nano_beir.arguana": 0.46876474079443375,
2310
+ "step": 3000
2311
+ },
2312
+ {
2313
+ "epoch": 2.1398002853067046,
2314
+ "eval/nano_beir.climate_fever": 0.24845534369031178,
2315
+ "step": 3000
2316
+ },
2317
+ {
2318
+ "epoch": 2.1398002853067046,
2319
+ "eval/nano_beir.dbpedia_entity": 0.3518699079588141,
2320
+ "step": 3000
2321
+ },
2322
+ {
2323
+ "epoch": 2.1398002853067046,
2324
+ "eval/nano_beir.fever": 0.5976831251710185,
2325
+ "step": 3000
2326
+ },
2327
+ {
2328
+ "epoch": 2.1398002853067046,
2329
+ "eval/nano_beir.fiqa": 0.40792762165216695,
2330
+ "step": 3000
2331
+ },
2332
+ {
2333
+ "epoch": 2.1398002853067046,
2334
+ "eval/nano_beir.hotpotqa": 0.6722148076317351,
2335
+ "step": 3000
2336
+ },
2337
+ {
2338
+ "epoch": 2.1398002853067046,
2339
+ "eval/nano_beir.nfcorpus": 0.34539657615301445,
2340
+ "step": 3000
2341
+ },
2342
+ {
2343
+ "epoch": 2.1398002853067046,
2344
+ "eval/nano_beir.nq": 0.38890377559243106,
2345
+ "step": 3000
2346
+ },
2347
+ {
2348
+ "epoch": 2.1398002853067046,
2349
+ "eval/nano_beir.quora": 0.826494799130688,
2350
+ "step": 3000
2351
+ },
2352
+ {
2353
+ "epoch": 2.1398002853067046,
2354
+ "eval/nano_beir.scidocs": 0.2518276219672329,
2355
+ "step": 3000
2356
+ },
2357
+ {
2358
+ "epoch": 2.1398002853067046,
2359
+ "eval/nano_beir.scifact": 0.7707161514794307,
2360
+ "step": 3000
2361
+ },
2362
+ {
2363
+ "epoch": 2.1398002853067046,
2364
+ "eval/nano_beir.webis_touche2020": 0.2854043806246762,
2365
+ "step": 3000
2366
+ },
2367
+ {
2368
+ "epoch": 2.1398002853067046,
2369
+ "eval/avg": 0.46797157098716285,
2370
+ "step": 3000
2371
+ },
2372
+ {
2373
+ "epoch": 2.1469329529243937,
2374
+ "grad_norm": 4.59375,
2375
+ "learning_rate": 4.100837805544279e-05,
2376
+ "loss": 0.7768,
2377
+ "step": 3010
2378
+ },
2379
+ {
2380
+ "epoch": 2.1540656205420827,
2381
+ "grad_norm": 5.0625,
2382
+ "learning_rate": 4.0665224844564053e-05,
2383
+ "loss": 0.6445,
2384
+ "step": 3020
2385
+ },
2386
+ {
2387
+ "epoch": 2.1611982881597718,
2388
+ "grad_norm": 4.5,
2389
+ "learning_rate": 4.032252659745699e-05,
2390
+ "loss": 0.7335,
2391
+ "step": 3030
2392
+ },
2393
+ {
2394
+ "epoch": 2.168330955777461,
2395
+ "grad_norm": 4.59375,
2396
+ "learning_rate": 3.9980300016750694e-05,
2397
+ "loss": 0.742,
2398
+ "step": 3040
2399
+ },
2400
+ {
2401
+ "epoch": 2.17546362339515,
2402
+ "grad_norm": 4.34375,
2403
+ "learning_rate": 3.963856178208588e-05,
2404
+ "loss": 0.7422,
2405
+ "step": 3050
2406
+ },
2407
+ {
2408
+ "epoch": 2.182596291012839,
2409
+ "grad_norm": 5.125,
2410
+ "learning_rate": 3.9297328549302e-05,
2411
+ "loss": 0.7333,
2412
+ "step": 3060
2413
+ },
2414
+ {
2415
+ "epoch": 2.189728958630528,
2416
+ "grad_norm": 4.5625,
2417
+ "learning_rate": 3.895661694962542e-05,
2418
+ "loss": 0.8071,
2419
+ "step": 3070
2420
+ },
2421
+ {
2422
+ "epoch": 2.196861626248217,
2423
+ "grad_norm": 5.5,
2424
+ "learning_rate": 3.8616443588858805e-05,
2425
+ "loss": 0.7712,
2426
+ "step": 3080
2427
+ },
2428
+ {
2429
+ "epoch": 2.2039942938659056,
2430
+ "grad_norm": 4.53125,
2431
+ "learning_rate": 3.827682504657187e-05,
2432
+ "loss": 0.6875,
2433
+ "step": 3090
2434
+ },
2435
+ {
2436
+ "epoch": 2.2111269614835947,
2437
+ "grad_norm": 5.0,
2438
+ "learning_rate": 3.7937777875293244e-05,
2439
+ "loss": 0.7984,
2440
+ "step": 3100
2441
+ },
2442
+ {
2443
+ "epoch": 2.2182596291012837,
2444
+ "grad_norm": 4.78125,
2445
+ "learning_rate": 3.759931859970374e-05,
2446
+ "loss": 0.7392,
2447
+ "step": 3110
2448
+ },
2449
+ {
2450
+ "epoch": 2.2253922967189728,
2451
+ "grad_norm": 4.375,
2452
+ "learning_rate": 3.72614637158309e-05,
2453
+ "loss": 0.796,
2454
+ "step": 3120
2455
+ },
2456
+ {
2457
+ "epoch": 2.232524964336662,
2458
+ "grad_norm": 4.75,
2459
+ "learning_rate": 3.692422969024516e-05,
2460
+ "loss": 0.7527,
2461
+ "step": 3130
2462
+ },
2463
+ {
2464
+ "epoch": 2.239657631954351,
2465
+ "grad_norm": 5.0625,
2466
+ "learning_rate": 3.6587632959257165e-05,
2467
+ "loss": 0.7767,
2468
+ "step": 3140
2469
+ },
2470
+ {
2471
+ "epoch": 2.24679029957204,
2472
+ "grad_norm": 5.21875,
2473
+ "learning_rate": 3.625168992811671e-05,
2474
+ "loss": 0.7154,
2475
+ "step": 3150
2476
+ },
2477
+ {
2478
+ "epoch": 2.253922967189729,
2479
+ "grad_norm": 5.5,
2480
+ "learning_rate": 3.591641697021317e-05,
2481
+ "loss": 0.7039,
2482
+ "step": 3160
2483
+ },
2484
+ {
2485
+ "epoch": 2.261055634807418,
2486
+ "grad_norm": 5.46875,
2487
+ "learning_rate": 3.558183042627755e-05,
2488
+ "loss": 0.7844,
2489
+ "step": 3170
2490
+ },
2491
+ {
2492
+ "epoch": 2.268188302425107,
2493
+ "grad_norm": 3.96875,
2494
+ "learning_rate": 3.524794660358593e-05,
2495
+ "loss": 0.7178,
2496
+ "step": 3180
2497
+ },
2498
+ {
2499
+ "epoch": 2.275320970042796,
2500
+ "grad_norm": 5.03125,
2501
+ "learning_rate": 3.491478177516484e-05,
2502
+ "loss": 0.8085,
2503
+ "step": 3190
2504
+ },
2505
+ {
2506
+ "epoch": 2.282453637660485,
2507
+ "grad_norm": 3.96875,
2508
+ "learning_rate": 3.4582352178997935e-05,
2509
+ "loss": 0.7032,
2510
+ "step": 3200
2511
+ },
2512
+ {
2513
+ "epoch": 2.289586305278174,
2514
+ "grad_norm": 4.1875,
2515
+ "learning_rate": 3.425067401723477e-05,
2516
+ "loss": 0.7552,
2517
+ "step": 3210
2518
+ },
2519
+ {
2520
+ "epoch": 2.2967189728958632,
2521
+ "grad_norm": 5.5625,
2522
+ "learning_rate": 3.3919763455401015e-05,
2523
+ "loss": 0.7586,
2524
+ "step": 3220
2525
+ },
2526
+ {
2527
+ "epoch": 2.3038516405135523,
2528
+ "grad_norm": 5.46875,
2529
+ "learning_rate": 3.358963662161062e-05,
2530
+ "loss": 0.8767,
2531
+ "step": 3230
2532
+ },
2533
+ {
2534
+ "epoch": 2.310984308131241,
2535
+ "grad_norm": 5.625,
2536
+ "learning_rate": 3.326030960577972e-05,
2537
+ "loss": 0.7965,
2538
+ "step": 3240
2539
+ },
2540
+ {
2541
+ "epoch": 2.31811697574893,
2542
+ "grad_norm": 5.03125,
2543
+ "learning_rate": 3.293179845884245e-05,
2544
+ "loss": 0.725,
2545
+ "step": 3250
2546
+ },
2547
+ {
2548
+ "epoch": 2.325249643366619,
2549
+ "grad_norm": 5.28125,
2550
+ "learning_rate": 3.2604119191968654e-05,
2551
+ "loss": 0.6246,
2552
+ "step": 3260
2553
+ },
2554
+ {
2555
+ "epoch": 2.332382310984308,
2556
+ "grad_norm": 4.65625,
2557
+ "learning_rate": 3.2277287775783525e-05,
2558
+ "loss": 0.7659,
2559
+ "step": 3270
2560
+ },
2561
+ {
2562
+ "epoch": 2.339514978601997,
2563
+ "grad_norm": 5.9375,
2564
+ "learning_rate": 3.195132013958918e-05,
2565
+ "loss": 0.7054,
2566
+ "step": 3280
2567
+ },
2568
+ {
2569
+ "epoch": 2.346647646219686,
2570
+ "grad_norm": 5.40625,
2571
+ "learning_rate": 3.162623217058834e-05,
2572
+ "loss": 0.7813,
2573
+ "step": 3290
2574
+ },
2575
+ {
2576
+ "epoch": 2.353780313837375,
2577
+ "grad_norm": 5.25,
2578
+ "learning_rate": 3.130203971310999e-05,
2579
+ "loss": 0.7004,
2580
+ "step": 3300
2581
+ },
2582
+ {
2583
+ "epoch": 2.3609129814550642,
2584
+ "grad_norm": 4.59375,
2585
+ "learning_rate": 3.097875856783713e-05,
2586
+ "loss": 0.7509,
2587
+ "step": 3310
2588
+ },
2589
+ {
2590
+ "epoch": 2.3680456490727533,
2591
+ "grad_norm": 5.5625,
2592
+ "learning_rate": 3.065640449103669e-05,
2593
+ "loss": 0.7423,
2594
+ "step": 3320
2595
+ },
2596
+ {
2597
+ "epoch": 2.3751783166904423,
2598
+ "grad_norm": 4.5,
2599
+ "learning_rate": 3.033499319379163e-05,
2600
+ "loss": 0.7703,
2601
+ "step": 3330
2602
+ },
2603
+ {
2604
+ "epoch": 2.3823109843081314,
2605
+ "grad_norm": 4.03125,
2606
+ "learning_rate": 3.001454034123512e-05,
2607
+ "loss": 0.7913,
2608
+ "step": 3340
2609
+ },
2610
+ {
2611
+ "epoch": 2.3894436519258204,
2612
+ "grad_norm": 5.46875,
2613
+ "learning_rate": 2.9695061551787108e-05,
2614
+ "loss": 0.7646,
2615
+ "step": 3350
2616
+ },
2617
+ {
2618
+ "epoch": 2.3965763195435095,
2619
+ "grad_norm": 4.78125,
2620
+ "learning_rate": 2.9376572396393048e-05,
2621
+ "loss": 0.751,
2622
+ "step": 3360
2623
+ },
2624
+ {
2625
+ "epoch": 2.403708987161198,
2626
+ "grad_norm": 4.84375,
2627
+ "learning_rate": 2.9059088397765088e-05,
2628
+ "loss": 0.7969,
2629
+ "step": 3370
2630
+ },
2631
+ {
2632
+ "epoch": 2.410841654778887,
2633
+ "grad_norm": 3.9375,
2634
+ "learning_rate": 2.874262502962537e-05,
2635
+ "loss": 0.7995,
2636
+ "step": 3380
2637
+ },
2638
+ {
2639
+ "epoch": 2.417974322396576,
2640
+ "grad_norm": 9.625,
2641
+ "learning_rate": 2.8427197715952047e-05,
2642
+ "loss": 0.7095,
2643
+ "step": 3390
2644
+ },
2645
+ {
2646
+ "epoch": 2.425106990014265,
2647
+ "grad_norm": 7.78125,
2648
+ "learning_rate": 2.811282183022736e-05,
2649
+ "loss": 0.8267,
2650
+ "step": 3400
2651
+ },
2652
+ {
2653
+ "epoch": 2.4322396576319543,
2654
+ "grad_norm": 4.5625,
2655
+ "learning_rate": 2.7799512694688468e-05,
2656
+ "loss": 0.7129,
2657
+ "step": 3410
2658
+ },
2659
+ {
2660
+ "epoch": 2.4393723252496433,
2661
+ "grad_norm": 6.5,
2662
+ "learning_rate": 2.7487285579580637e-05,
2663
+ "loss": 0.8285,
2664
+ "step": 3420
2665
+ },
2666
+ {
2667
+ "epoch": 2.4465049928673324,
2668
+ "grad_norm": 5.09375,
2669
+ "learning_rate": 2.7176155702412937e-05,
2670
+ "loss": 0.7694,
2671
+ "step": 3430
2672
+ },
2673
+ {
2674
+ "epoch": 2.4536376604850214,
2675
+ "grad_norm": 4.34375,
2676
+ "learning_rate": 2.686613822721666e-05,
2677
+ "loss": 0.7047,
2678
+ "step": 3440
2679
+ },
2680
+ {
2681
+ "epoch": 2.4607703281027105,
2682
+ "grad_norm": 5.1875,
2683
+ "learning_rate": 2.6557248263806174e-05,
2684
+ "loss": 0.7772,
2685
+ "step": 3450
2686
+ },
2687
+ {
2688
+ "epoch": 2.4679029957203995,
2689
+ "grad_norm": 4.6875,
2690
+ "learning_rate": 2.6249500867042522e-05,
2691
+ "loss": 0.8148,
2692
+ "step": 3460
2693
+ },
2694
+ {
2695
+ "epoch": 2.4750356633380886,
2696
+ "grad_norm": 5.3125,
2697
+ "learning_rate": 2.5942911036099655e-05,
2698
+ "loss": 0.689,
2699
+ "step": 3470
2700
+ },
2701
+ {
2702
+ "epoch": 2.4821683309557776,
2703
+ "grad_norm": 5.21875,
2704
+ "learning_rate": 2.5637493713733374e-05,
2705
+ "loss": 0.7762,
2706
+ "step": 3480
2707
+ },
2708
+ {
2709
+ "epoch": 2.4893009985734667,
2710
+ "grad_norm": 5.59375,
2711
+ "learning_rate": 2.533326378555314e-05,
2712
+ "loss": 0.7899,
2713
+ "step": 3490
2714
+ },
2715
+ {
2716
+ "epoch": 2.4964336661911553,
2717
+ "grad_norm": 4.53125,
2718
+ "learning_rate": 2.5030236079296444e-05,
2719
+ "loss": 0.747,
2720
+ "step": 3500
2721
+ },
2722
+ {
2723
+ "epoch": 2.5035663338088447,
2724
+ "grad_norm": 4.125,
2725
+ "learning_rate": 2.4728425364106135e-05,
2726
+ "loss": 0.7218,
2727
+ "step": 3510
2728
+ },
2729
+ {
2730
+ "epoch": 2.5106990014265333,
2731
+ "grad_norm": 4.96875,
2732
+ "learning_rate": 2.442784634981071e-05,
2733
+ "loss": 0.7172,
2734
+ "step": 3520
2735
+ },
2736
+ {
2737
+ "epoch": 2.5178316690442224,
2738
+ "grad_norm": 4.40625,
2739
+ "learning_rate": 2.412851368620726e-05,
2740
+ "loss": 0.6116,
2741
+ "step": 3530
2742
+ },
2743
+ {
2744
+ "epoch": 2.5249643366619114,
2745
+ "grad_norm": 4.4375,
2746
+ "learning_rate": 2.383044196234753e-05,
2747
+ "loss": 0.7398,
2748
+ "step": 3540
2749
+ },
2750
+ {
2751
+ "epoch": 2.5320970042796005,
2752
+ "grad_norm": 5.09375,
2753
+ "learning_rate": 2.3533645705826806e-05,
2754
+ "loss": 0.8014,
2755
+ "step": 3550
2756
+ },
2757
+ {
2758
+ "epoch": 2.5392296718972895,
2759
+ "grad_norm": 5.1875,
2760
+ "learning_rate": 2.323813938207593e-05,
2761
+ "loss": 0.775,
2762
+ "step": 3560
2763
+ },
2764
+ {
2765
+ "epoch": 2.5463623395149786,
2766
+ "grad_norm": 4.28125,
2767
+ "learning_rate": 2.294393739365621e-05,
2768
+ "loss": 0.8234,
2769
+ "step": 3570
2770
+ },
2771
+ {
2772
+ "epoch": 2.5534950071326676,
2773
+ "grad_norm": 4.9375,
2774
+ "learning_rate": 2.265105407955752e-05,
2775
+ "loss": 0.7388,
2776
+ "step": 3580
2777
+ },
2778
+ {
2779
+ "epoch": 2.5606276747503567,
2780
+ "grad_norm": 5.125,
2781
+ "learning_rate": 2.235950371449938e-05,
2782
+ "loss": 0.6721,
2783
+ "step": 3590
2784
+ },
2785
+ {
2786
+ "epoch": 2.5677603423680457,
2787
+ "grad_norm": 5.3125,
2788
+ "learning_rate": 2.2069300508235275e-05,
2789
+ "loss": 0.7251,
2790
+ "step": 3600
2791
+ },
2792
+ {
2793
+ "epoch": 2.574893009985735,
2794
+ "grad_norm": 6.46875,
2795
+ "learning_rate": 2.1780458604860056e-05,
2796
+ "loss": 0.7803,
2797
+ "step": 3610
2798
+ },
2799
+ {
2800
+ "epoch": 2.582025677603424,
2801
+ "grad_norm": 4.78125,
2802
+ "learning_rate": 2.14929920821206e-05,
2803
+ "loss": 0.7447,
2804
+ "step": 3620
2805
+ },
2806
+ {
2807
+ "epoch": 2.5891583452211124,
2808
+ "grad_norm": 5.15625,
2809
+ "learning_rate": 2.1206914950729672e-05,
2810
+ "loss": 0.7869,
2811
+ "step": 3630
2812
+ },
2813
+ {
2814
+ "epoch": 2.596291012838802,
2815
+ "grad_norm": 3.46875,
2816
+ "learning_rate": 2.0922241153683066e-05,
2817
+ "loss": 0.7073,
2818
+ "step": 3640
2819
+ },
2820
+ {
2821
+ "epoch": 2.6034236804564905,
2822
+ "grad_norm": 5.125,
2823
+ "learning_rate": 2.0638984565580023e-05,
2824
+ "loss": 0.7132,
2825
+ "step": 3650
2826
+ },
2827
+ {
2828
+ "epoch": 2.6105563480741796,
2829
+ "grad_norm": 4.875,
2830
+ "learning_rate": 2.035715899194704e-05,
2831
+ "loss": 0.7917,
2832
+ "step": 3660
2833
+ },
2834
+ {
2835
+ "epoch": 2.6176890156918686,
2836
+ "grad_norm": 4.84375,
2837
+ "learning_rate": 2.007677816856498e-05,
2838
+ "loss": 0.7058,
2839
+ "step": 3670
2840
+ },
2841
+ {
2842
+ "epoch": 2.6248216833095577,
2843
+ "grad_norm": 4.65625,
2844
+ "learning_rate": 1.9797855760799612e-05,
2845
+ "loss": 0.7053,
2846
+ "step": 3680
2847
+ },
2848
+ {
2849
+ "epoch": 2.6319543509272467,
2850
+ "grad_norm": 5.65625,
2851
+ "learning_rate": 1.9520405362935594e-05,
2852
+ "loss": 0.7725,
2853
+ "step": 3690
2854
+ },
2855
+ {
2856
+ "epoch": 2.6390870185449358,
2857
+ "grad_norm": 4.34375,
2858
+ "learning_rate": 1.9244440497513893e-05,
2859
+ "loss": 0.712,
2860
+ "step": 3700
2861
+ },
2862
+ {
2863
+ "epoch": 2.646219686162625,
2864
+ "grad_norm": 5.0625,
2865
+ "learning_rate": 1.896997461467272e-05,
2866
+ "loss": 0.7954,
2867
+ "step": 3710
2868
+ },
2869
+ {
2870
+ "epoch": 2.653352353780314,
2871
+ "grad_norm": 5.0625,
2872
+ "learning_rate": 1.869702109149199e-05,
2873
+ "loss": 0.727,
2874
+ "step": 3720
2875
+ },
2876
+ {
2877
+ "epoch": 2.660485021398003,
2878
+ "grad_norm": 3.78125,
2879
+ "learning_rate": 1.842559323134136e-05,
2880
+ "loss": 0.6694,
2881
+ "step": 3730
2882
+ },
2883
+ {
2884
+ "epoch": 2.667617689015692,
2885
+ "grad_norm": 5.15625,
2886
+ "learning_rate": 1.8155704263231776e-05,
2887
+ "loss": 0.7812,
2888
+ "step": 3740
2889
+ },
2890
+ {
2891
+ "epoch": 2.674750356633381,
2892
+ "grad_norm": 5.25,
2893
+ "learning_rate": 1.788736734117078e-05,
2894
+ "loss": 0.756,
2895
+ "step": 3750
2896
+ },
2897
+ {
2898
+ "epoch": 2.68188302425107,
2899
+ "grad_norm": 4.46875,
2900
+ "learning_rate": 1.762059554352143e-05,
2901
+ "loss": 0.6954,
2902
+ "step": 3760
2903
+ },
2904
+ {
2905
+ "epoch": 2.689015691868759,
2906
+ "grad_norm": 4.84375,
2907
+ "learning_rate": 1.7355401872364758e-05,
2908
+ "loss": 0.7037,
2909
+ "step": 3770
2910
+ },
2911
+ {
2912
+ "epoch": 2.6961483594864477,
2913
+ "grad_norm": 5.65625,
2914
+ "learning_rate": 1.709179925286617e-05,
2915
+ "loss": 0.7861,
2916
+ "step": 3780
2917
+ },
2918
+ {
2919
+ "epoch": 2.703281027104137,
2920
+ "grad_norm": 4.65625,
2921
+ "learning_rate": 1.6829800532645447e-05,
2922
+ "loss": 0.7455,
2923
+ "step": 3790
2924
+ },
2925
+ {
2926
+ "epoch": 2.710413694721826,
2927
+ "grad_norm": 5.375,
2928
+ "learning_rate": 1.6569418481150595e-05,
2929
+ "loss": 0.6718,
2930
+ "step": 3800
2931
+ },
2932
+ {
2933
+ "epoch": 2.717546362339515,
2934
+ "grad_norm": 5.09375,
2935
+ "learning_rate": 1.6310665789035467e-05,
2936
+ "loss": 0.7419,
2937
+ "step": 3810
2938
+ },
2939
+ {
2940
+ "epoch": 2.724679029957204,
2941
+ "grad_norm": 5.8125,
2942
+ "learning_rate": 1.605355506754121e-05,
2943
+ "loss": 0.7336,
2944
+ "step": 3820
2945
+ },
2946
+ {
2947
+ "epoch": 2.731811697574893,
2948
+ "grad_norm": 7.40625,
2949
+ "learning_rate": 1.5798098847881666e-05,
2950
+ "loss": 0.6514,
2951
+ "step": 3830
2952
+ },
2953
+ {
2954
+ "epoch": 2.738944365192582,
2955
+ "grad_norm": 4.25,
2956
+ "learning_rate": 1.554430958063259e-05,
2957
+ "loss": 0.6845,
2958
+ "step": 3840
2959
+ },
2960
+ {
2961
+ "epoch": 2.746077032810271,
2962
+ "grad_norm": 4.96875,
2963
+ "learning_rate": 1.529219963512481e-05,
2964
+ "loss": 0.7397,
2965
+ "step": 3850
2966
+ },
2967
+ {
2968
+ "epoch": 2.75320970042796,
2969
+ "grad_norm": 3.875,
2970
+ "learning_rate": 1.5041781298841423e-05,
2971
+ "loss": 0.7563,
2972
+ "step": 3860
2973
+ },
2974
+ {
2975
+ "epoch": 2.760342368045649,
2976
+ "grad_norm": 4.875,
2977
+ "learning_rate": 1.4793066776818842e-05,
2978
+ "loss": 0.7316,
2979
+ "step": 3870
2980
+ },
2981
+ {
2982
+ "epoch": 2.767475035663338,
2983
+ "grad_norm": 4.125,
2984
+ "learning_rate": 1.4546068191051987e-05,
2985
+ "loss": 0.7016,
2986
+ "step": 3880
2987
+ },
2988
+ {
2989
+ "epoch": 2.7746077032810272,
2990
+ "grad_norm": 5.4375,
2991
+ "learning_rate": 1.4300797579903474e-05,
2992
+ "loss": 0.7698,
2993
+ "step": 3890
2994
+ },
2995
+ {
2996
+ "epoch": 2.7817403708987163,
2997
+ "grad_norm": 5.625,
2998
+ "learning_rate": 1.4057266897516841e-05,
2999
+ "loss": 0.8646,
3000
+ "step": 3900
3001
+ },
3002
+ {
3003
+ "epoch": 2.788873038516405,
3004
+ "grad_norm": 5.8125,
3005
+ "learning_rate": 1.3815488013233984e-05,
3006
+ "loss": 0.7541,
3007
+ "step": 3910
3008
+ },
3009
+ {
3010
+ "epoch": 2.7960057061340944,
3011
+ "grad_norm": 5.6875,
3012
+ "learning_rate": 1.3575472711016635e-05,
3013
+ "loss": 0.8134,
3014
+ "step": 3920
3015
+ },
3016
+ {
3017
+ "epoch": 2.803138373751783,
3018
+ "grad_norm": 5.0,
3019
+ "learning_rate": 1.3337232688872009e-05,
3020
+ "loss": 0.7447,
3021
+ "step": 3930
3022
+ },
3023
+ {
3024
+ "epoch": 2.810271041369472,
3025
+ "grad_norm": 4.46875,
3026
+ "learning_rate": 1.3100779558282672e-05,
3027
+ "loss": 0.6445,
3028
+ "step": 3940
3029
+ },
3030
+ {
3031
+ "epoch": 2.817403708987161,
3032
+ "grad_norm": 4.46875,
3033
+ "learning_rate": 1.2866124843640615e-05,
3034
+ "loss": 0.7147,
3035
+ "step": 3950
3036
+ },
3037
+ {
3038
+ "epoch": 2.82453637660485,
3039
+ "grad_norm": 5.0,
3040
+ "learning_rate": 1.2633279981685608e-05,
3041
+ "loss": 0.7744,
3042
+ "step": 3960
3043
+ },
3044
+ {
3045
+ "epoch": 2.831669044222539,
3046
+ "grad_norm": 4.03125,
3047
+ "learning_rate": 1.240225632094773e-05,
3048
+ "loss": 0.7242,
3049
+ "step": 3970
3050
+ },
3051
+ {
3052
+ "epoch": 2.8388017118402282,
3053
+ "grad_norm": 4.6875,
3054
+ "learning_rate": 1.217306512119425e-05,
3055
+ "loss": 0.7314,
3056
+ "step": 3980
3057
+ },
3058
+ {
3059
+ "epoch": 2.8459343794579173,
3060
+ "grad_norm": 4.96875,
3061
+ "learning_rate": 1.194571755288092e-05,
3062
+ "loss": 0.7505,
3063
+ "step": 3990
3064
+ },
3065
+ {
3066
+ "epoch": 2.8530670470756063,
3067
+ "grad_norm": 5.1875,
3068
+ "learning_rate": 1.1720224696607474e-05,
3069
+ "loss": 0.7146,
3070
+ "step": 4000
3071
+ }
3072
+ ],
3073
+ "logging_steps": 10,
3074
+ "max_steps": 5000,
3075
+ "num_input_tokens_seen": 0,
3076
+ "num_train_epochs": 4,
3077
+ "save_steps": 1000,
3078
+ "stateful_callbacks": {
3079
+ "TrainerControl": {
3080
+ "args": {
3081
+ "should_epoch_stop": false,
3082
+ "should_evaluate": false,
3083
+ "should_log": false,
3084
+ "should_save": true,
3085
+ "should_training_stop": false
3086
+ },
3087
+ "attributes": {}
3088
+ }
3089
+ },
3090
+ "total_flos": 0.0,
3091
+ "train_batch_size": 16,
3092
+ "trial_name": null,
3093
+ "trial_params": null
3094
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-4000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af7cf7c3c741add5e8996639d61d87fe74bb40ad0faa1b29901130f4b3a60fc3
3
+ size 6353
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/checkpoint-5000/moved ADDED
File without changes
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "embedding_dropout": 0.0,
16
+ "eos_token_id": 50282,
17
+ "global_attn_every_n_layers": 3,
18
+ "global_rope_theta": 160000.0,
19
+ "gradient_checkpointing": false,
20
+ "hidden_activation": "gelu",
21
+ "hidden_size": 768,
22
+ "initializer_cutoff_factor": 2.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 1152,
25
+ "layer_norm_eps": 1e-05,
26
+ "local_attention": 128,
27
+ "local_rope_theta": 10000.0,
28
+ "max_position_embeddings": 8192,
29
+ "mlp_bias": false,
30
+ "mlp_dropout": 0.0,
31
+ "model_type": "modernbert",
32
+ "norm_bias": false,
33
+ "norm_eps": 1e-05,
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 22,
36
+ "pad_token_id": 50283,
37
+ "position_embedding_type": "absolute",
38
+ "repad_logits_with_grad": false,
39
+ "sep_token_id": 50282,
40
+ "sparse_pred_ignore_index": -100,
41
+ "sparse_prediction": false,
42
+ "torch_dtype": "bfloat16",
43
+ "transformers_version": "4.53.0",
44
+ "vocab_size": 50368
45
+ }
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b8cee98ee2a3b201eac58f67d136f3636e25ea77f2d6456aed0e76ba9806ba7
3
+ size 298041696
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:466c5a0c45557ab3167c47df6fc8d83c9a350c6d83e31f3f63b06650c8ebae6e
3
+ size 596170443
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24c12cceaffe8acec04189ba71dbb81f100dd5cd03b86390253117b3fab2911d
3
+ size 15365
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69adbd9997461473344beb9c44d2e496e24fbc4d6fe69245ab0bd127882efd96
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baaf7026d9e588ca2ba5b4de8768a379982b7530f39e0b88fa44af9bef8e8bc9
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ec81cbdf599c90b49fb13998f6e46e2492b55345216231b6ce078f88cf04eae
3
+ size 15429
modernbert-crux-researchy-pos_20.neg_51.b64_n512.1e-4.request/runs/Jan01_22-56-07_nid005051/events.out.tfevents.1767301064.nid005051.19581.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f851e7139ea8140e18315ef864967a2ff4a57ec78684c6323cb8173fc14f535
3
+ size 8790