irishprancer commited on
Commit
2099f59
·
verified ·
1 Parent(s): ea7e057

Training in progress, step 750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c85ccd2d8f7fed25fa90e9cbc1809dc3678b16093649b77b9d0294a5d7521fc
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77f5c4bb61551feebcac9c0da6b303e325c053a4fd69c4f093ef1a5cdd526908
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12713859786a4771f1f53038cc506f8a2af69a22a568239e40fab40d3c479ed8
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fb0635ad14c320dca7f97c2d3d371f8ca3e9678b324e2f75556c62bccc96c5c
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2616293446c1c2f2c6b0e270c4bb50dd25782b075145433b30550e3bb3a1845c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30b57ec052aeaf035e39cd5057ce6977dfdadd2be87b9e8aaa1d9afd55d2f8e1
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da0b9a1e2fa3da24e9af8b74787d70ff4c95d9bc9b74eeab135df0350a00b462
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8974fc0327a603be8a90ee235d3ef81c4ee4af2818c873bdeb9bfb38f108e1a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7165840268135071,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
- "epoch": 19.565217391304348,
5
  "eval_steps": 150,
6
- "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -346,6 +346,232 @@
346
  "eval_samples_per_second": 24.844,
347
  "eval_steps_per_second": 24.844,
348
  "step": 450
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
349
  }
350
  ],
351
  "logging_steps": 10,
@@ -365,7 +591,7 @@
365
  "attributes": {}
366
  }
367
  },
368
- "total_flos": 1.1591800814075904e+16,
369
  "train_batch_size": 4,
370
  "trial_name": null,
371
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7165840268135071,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
+ "epoch": 32.608695652173914,
5
  "eval_steps": 150,
6
+ "global_step": 750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
346
  "eval_samples_per_second": 24.844,
347
  "eval_steps_per_second": 24.844,
348
  "step": 450
349
+ },
350
+ {
351
+ "epoch": 20.0,
352
+ "grad_norm": 2.1217708587646484,
353
+ "learning_rate": 2.9996147467351856e-05,
354
+ "loss": 0.5145,
355
+ "step": 460
356
+ },
357
+ {
358
+ "epoch": 20.434782608695652,
359
+ "grad_norm": 1.275344967842102,
360
+ "learning_rate": 2.9995930474939773e-05,
361
+ "loss": 0.4782,
362
+ "step": 470
363
+ },
364
+ {
365
+ "epoch": 20.869565217391305,
366
+ "grad_norm": 1.4735126495361328,
367
+ "learning_rate": 2.9995707538619975e-05,
368
+ "loss": 0.5707,
369
+ "step": 480
370
+ },
371
+ {
372
+ "epoch": 21.304347826086957,
373
+ "grad_norm": 1.32221257686615,
374
+ "learning_rate": 2.9995478658480822e-05,
375
+ "loss": 0.5165,
376
+ "step": 490
377
+ },
378
+ {
379
+ "epoch": 21.73913043478261,
380
+ "grad_norm": 1.2421287298202515,
381
+ "learning_rate": 2.9995243834613043e-05,
382
+ "loss": 0.5205,
383
+ "step": 500
384
+ },
385
+ {
386
+ "epoch": 22.17391304347826,
387
+ "grad_norm": 1.7849466800689697,
388
+ "learning_rate": 2.9995003067109707e-05,
389
+ "loss": 0.4834,
390
+ "step": 510
391
+ },
392
+ {
393
+ "epoch": 22.608695652173914,
394
+ "grad_norm": 1.5369889736175537,
395
+ "learning_rate": 2.9994756356066246e-05,
396
+ "loss": 0.5613,
397
+ "step": 520
398
+ },
399
+ {
400
+ "epoch": 23.043478260869566,
401
+ "grad_norm": 1.7313120365142822,
402
+ "learning_rate": 2.999450370158046e-05,
403
+ "loss": 0.4928,
404
+ "step": 530
405
+ },
406
+ {
407
+ "epoch": 23.47826086956522,
408
+ "grad_norm": 1.3083937168121338,
409
+ "learning_rate": 2.9994245103752478e-05,
410
+ "loss": 0.4382,
411
+ "step": 540
412
+ },
413
+ {
414
+ "epoch": 23.91304347826087,
415
+ "grad_norm": 1.233344554901123,
416
+ "learning_rate": 2.999398056268481e-05,
417
+ "loss": 0.5264,
418
+ "step": 550
419
+ },
420
+ {
421
+ "epoch": 24.347826086956523,
422
+ "grad_norm": 1.4090428352355957,
423
+ "learning_rate": 2.9993710078482306e-05,
424
+ "loss": 0.5206,
425
+ "step": 560
426
+ },
427
+ {
428
+ "epoch": 24.782608695652176,
429
+ "grad_norm": 0.9520849585533142,
430
+ "learning_rate": 2.9993433651252185e-05,
431
+ "loss": 0.4429,
432
+ "step": 570
433
+ },
434
+ {
435
+ "epoch": 25.217391304347824,
436
+ "grad_norm": 1.7179316282272339,
437
+ "learning_rate": 2.9993151281104006e-05,
438
+ "loss": 0.5329,
439
+ "step": 580
440
+ },
441
+ {
442
+ "epoch": 25.652173913043477,
443
+ "grad_norm": 1.1341142654418945,
444
+ "learning_rate": 2.9992862968149695e-05,
445
+ "loss": 0.4735,
446
+ "step": 590
447
+ },
448
+ {
449
+ "epoch": 26.08695652173913,
450
+ "grad_norm": 1.1748226881027222,
451
+ "learning_rate": 2.9992568712503533e-05,
452
+ "loss": 0.4608,
453
+ "step": 600
454
+ },
455
+ {
456
+ "epoch": 26.08695652173913,
457
+ "eval_loss": 0.7199766039848328,
458
+ "eval_runtime": 0.5422,
459
+ "eval_samples_per_second": 18.442,
460
+ "eval_steps_per_second": 18.442,
461
+ "step": 600
462
+ },
463
+ {
464
+ "epoch": 26.52173913043478,
465
+ "grad_norm": 1.0799683332443237,
466
+ "learning_rate": 2.9992268514282142e-05,
467
+ "loss": 0.5121,
468
+ "step": 610
469
+ },
470
+ {
471
+ "epoch": 26.956521739130434,
472
+ "grad_norm": 1.3288029432296753,
473
+ "learning_rate": 2.999196237360452e-05,
474
+ "loss": 0.4316,
475
+ "step": 620
476
+ },
477
+ {
478
+ "epoch": 27.391304347826086,
479
+ "grad_norm": 1.1912176609039307,
480
+ "learning_rate": 2.9991650290592016e-05,
481
+ "loss": 0.4763,
482
+ "step": 630
483
+ },
484
+ {
485
+ "epoch": 27.82608695652174,
486
+ "grad_norm": 1.3694884777069092,
487
+ "learning_rate": 2.999133226536832e-05,
488
+ "loss": 0.5015,
489
+ "step": 640
490
+ },
491
+ {
492
+ "epoch": 28.26086956521739,
493
+ "grad_norm": 1.3749481439590454,
494
+ "learning_rate": 2.9991008298059493e-05,
495
+ "loss": 0.4116,
496
+ "step": 650
497
+ },
498
+ {
499
+ "epoch": 28.695652173913043,
500
+ "grad_norm": 1.5401527881622314,
501
+ "learning_rate": 2.9990678388793944e-05,
502
+ "loss": 0.5073,
503
+ "step": 660
504
+ },
505
+ {
506
+ "epoch": 29.130434782608695,
507
+ "grad_norm": 1.2782667875289917,
508
+ "learning_rate": 2.999034253770244e-05,
509
+ "loss": 0.4355,
510
+ "step": 670
511
+ },
512
+ {
513
+ "epoch": 29.565217391304348,
514
+ "grad_norm": 1.0564625263214111,
515
+ "learning_rate": 2.9990000744918097e-05,
516
+ "loss": 0.4714,
517
+ "step": 680
518
+ },
519
+ {
520
+ "epoch": 30.0,
521
+ "grad_norm": 2.43731427192688,
522
+ "learning_rate": 2.9989653010576392e-05,
523
+ "loss": 0.4153,
524
+ "step": 690
525
+ },
526
+ {
527
+ "epoch": 30.434782608695652,
528
+ "grad_norm": 1.2880218029022217,
529
+ "learning_rate": 2.9989299334815158e-05,
530
+ "loss": 0.4778,
531
+ "step": 700
532
+ },
533
+ {
534
+ "epoch": 30.869565217391305,
535
+ "grad_norm": 1.5784273147583008,
536
+ "learning_rate": 2.9988939717774578e-05,
537
+ "loss": 0.4135,
538
+ "step": 710
539
+ },
540
+ {
541
+ "epoch": 31.304347826086957,
542
+ "grad_norm": 0.8963285684585571,
543
+ "learning_rate": 2.9988574159597194e-05,
544
+ "loss": 0.4262,
545
+ "step": 720
546
+ },
547
+ {
548
+ "epoch": 31.73913043478261,
549
+ "grad_norm": 1.6010137796401978,
550
+ "learning_rate": 2.9988202660427907e-05,
551
+ "loss": 0.4843,
552
+ "step": 730
553
+ },
554
+ {
555
+ "epoch": 32.17391304347826,
556
+ "grad_norm": 1.1745761632919312,
557
+ "learning_rate": 2.9987825220413958e-05,
558
+ "loss": 0.4404,
559
+ "step": 740
560
+ },
561
+ {
562
+ "epoch": 32.608695652173914,
563
+ "grad_norm": 1.7309991121292114,
564
+ "learning_rate": 2.998744183970496e-05,
565
+ "loss": 0.4753,
566
+ "step": 750
567
+ },
568
+ {
569
+ "epoch": 32.608695652173914,
570
+ "eval_loss": 0.7314043045043945,
571
+ "eval_runtime": 0.4014,
572
+ "eval_samples_per_second": 24.911,
573
+ "eval_steps_per_second": 24.911,
574
+ "step": 750
575
  }
576
  ],
577
  "logging_steps": 10,
 
591
  "attributes": {}
592
  }
593
  },
594
+ "total_flos": 1.9231766387195904e+16,
595
  "train_batch_size": 4,
596
  "trial_name": null,
597
  "trial_params": null