aghatage commited on
Commit
f067d3e
·
verified ·
1 Parent(s): 54100dd

Training in progress, step 6500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18b75b4abf894259d125060201a5ccf51810364d91fd3cea0d60d17a6403b9f6
3
  size 12017472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b37eb9bcf96214de200298cdc8128de031674bd22404378d2eb8711f7a1f0639
3
  size 12017472
last-checkpoint/global_step6500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de937880dd5d112efb485014c98ed5da078bb734ff16262f3b7d295b12c4511a
3
+ size 71982309
last-checkpoint/global_step6500/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a6f69d58549c317558501ef11c494883a7fbe070adb7912f9014823b8c0cc54
3
+ size 146356645
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step6000
 
1
+ global_step6500
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8941841568d79a9d2dc41b93071be9e47935d7e9eda669b8a7caaabc7faae599
3
  size 14709
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fba9e07381b5475da97b7a36fb2ebd7e3dbf5fc84575cf7978cdcac8ad340c8
3
  size 14709
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 6000,
3
- "best_metric": 0.578772783279419,
4
- "best_model_checkpoint": "/root/leap-finetune/outputs/sft/lfm2_350m_marathi_optimized_12ep/checkpoint-6000",
5
- "epoch": 4.360661697873114,
6
  "eval_steps": 250,
7
- "global_step": 6000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2417,6 +2417,206 @@
2417
  "eval_samples_per_second": 43.383,
2418
  "eval_steps_per_second": 5.43,
2419
  "step": 6000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2420
  }
2421
  ],
2422
  "logging_steps": 25,
@@ -2436,7 +2636,7 @@
2436
  "attributes": {}
2437
  }
2438
  },
2439
- "total_flos": 3.3318287854639514e+17,
2440
  "train_batch_size": 4,
2441
  "trial_name": null,
2442
  "trial_params": null
 
1
  {
2
+ "best_global_step": 6500,
3
+ "best_metric": 0.5727137923240662,
4
+ "best_model_checkpoint": "/root/leap-finetune/outputs/sft/lfm2_350m_marathi_optimized_12ep/checkpoint-6500",
5
+ "epoch": 4.72423195782585,
6
  "eval_steps": 250,
7
+ "global_step": 6500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2417
  "eval_samples_per_second": 43.383,
2418
  "eval_steps_per_second": 5.43,
2419
  "step": 6000
2420
+ },
2421
+ {
2422
+ "epoch": 4.3788402108707505,
2423
+ "grad_norm": 0.7740823030471802,
2424
+ "learning_rate": 5.6915524395595964e-05,
2425
+ "loss": 0.5768,
2426
+ "mean_token_accuracy": 0.820943498313427,
2427
+ "num_tokens": 132643717.0,
2428
+ "step": 6025
2429
+ },
2430
+ {
2431
+ "epoch": 4.3970187238683875,
2432
+ "grad_norm": 0.7765363454818726,
2433
+ "learning_rate": 5.674186960500381e-05,
2434
+ "loss": 0.5673,
2435
+ "mean_token_accuracy": 0.8246706813573838,
2436
+ "num_tokens": 133190852.0,
2437
+ "step": 6050
2438
+ },
2439
+ {
2440
+ "epoch": 4.415197236866025,
2441
+ "grad_norm": 0.7921891808509827,
2442
+ "learning_rate": 5.656783140755755e-05,
2443
+ "loss": 0.5745,
2444
+ "mean_token_accuracy": 0.8214689978957176,
2445
+ "num_tokens": 133742448.0,
2446
+ "step": 6075
2447
+ },
2448
+ {
2449
+ "epoch": 4.433375749863661,
2450
+ "grad_norm": 0.7477549910545349,
2451
+ "learning_rate": 5.639341378891948e-05,
2452
+ "loss": 0.5649,
2453
+ "mean_token_accuracy": 0.8251917722821236,
2454
+ "num_tokens": 134278075.0,
2455
+ "step": 6100
2456
+ },
2457
+ {
2458
+ "epoch": 4.451554262861298,
2459
+ "grad_norm": 0.8201608657836914,
2460
+ "learning_rate": 5.6218620743440986e-05,
2461
+ "loss": 0.5688,
2462
+ "mean_token_accuracy": 0.8233516532182693,
2463
+ "num_tokens": 134830841.0,
2464
+ "step": 6125
2465
+ },
2466
+ {
2467
+ "epoch": 4.469732775858935,
2468
+ "grad_norm": 0.7502488493919373,
2469
+ "learning_rate": 5.6043456274071164e-05,
2470
+ "loss": 0.5667,
2471
+ "mean_token_accuracy": 0.824826996922493,
2472
+ "num_tokens": 135383251.0,
2473
+ "step": 6150
2474
+ },
2475
+ {
2476
+ "epoch": 4.487911288856571,
2477
+ "grad_norm": 0.7607132196426392,
2478
+ "learning_rate": 5.586792439226512e-05,
2479
+ "loss": 0.5634,
2480
+ "mean_token_accuracy": 0.8251435077190399,
2481
+ "num_tokens": 135935805.0,
2482
+ "step": 6175
2483
+ },
2484
+ {
2485
+ "epoch": 4.506089801854208,
2486
+ "grad_norm": 0.7952844500541687,
2487
+ "learning_rate": 5.569202911789206e-05,
2488
+ "loss": 0.566,
2489
+ "mean_token_accuracy": 0.8234961378574371,
2490
+ "num_tokens": 136496942.0,
2491
+ "step": 6200
2492
+ },
2493
+ {
2494
+ "epoch": 4.524268314851845,
2495
+ "grad_norm": 0.7817985415458679,
2496
+ "learning_rate": 5.551577447914333e-05,
2497
+ "loss": 0.5724,
2498
+ "mean_token_accuracy": 0.8220775499939919,
2499
+ "num_tokens": 137049699.0,
2500
+ "step": 6225
2501
+ },
2502
+ {
2503
+ "epoch": 4.542446827849482,
2504
+ "grad_norm": 0.8044058680534363,
2505
+ "learning_rate": 5.533916451244008e-05,
2506
+ "loss": 0.5707,
2507
+ "mean_token_accuracy": 0.8230817872285843,
2508
+ "num_tokens": 137612608.0,
2509
+ "step": 6250
2510
+ },
2511
+ {
2512
+ "epoch": 4.542446827849482,
2513
+ "eval_loss": 0.5755720734596252,
2514
+ "eval_mean_token_accuracy": 0.8204464651400747,
2515
+ "eval_num_tokens": 137612608.0,
2516
+ "eval_runtime": 114.4855,
2517
+ "eval_samples_per_second": 42.713,
2518
+ "eval_steps_per_second": 5.346,
2519
+ "step": 6250
2520
+ },
2521
+ {
2522
+ "epoch": 4.5606253408471185,
2523
+ "grad_norm": 0.7777162194252014,
2524
+ "learning_rate": 5.516220326234084e-05,
2525
+ "loss": 0.5557,
2526
+ "mean_token_accuracy": 0.8273163467645646,
2527
+ "num_tokens": 138155250.0,
2528
+ "step": 6275
2529
+ },
2530
+ {
2531
+ "epoch": 4.5788038538447555,
2532
+ "grad_norm": 0.7649146318435669,
2533
+ "learning_rate": 5.4984894781448934e-05,
2534
+ "loss": 0.5647,
2535
+ "mean_token_accuracy": 0.8240767633914947,
2536
+ "num_tokens": 138700597.0,
2537
+ "step": 6300
2538
+ },
2539
+ {
2540
+ "epoch": 4.596982366842393,
2541
+ "grad_norm": 0.7625702023506165,
2542
+ "learning_rate": 5.480724313031962e-05,
2543
+ "loss": 0.5638,
2544
+ "mean_token_accuracy": 0.8254871436953545,
2545
+ "num_tokens": 139240284.0,
2546
+ "step": 6325
2547
+ },
2548
+ {
2549
+ "epoch": 4.615160879840029,
2550
+ "grad_norm": 0.7333512306213379,
2551
+ "learning_rate": 5.462925237736715e-05,
2552
+ "loss": 0.5753,
2553
+ "mean_token_accuracy": 0.8208840799331665,
2554
+ "num_tokens": 139806757.0,
2555
+ "step": 6350
2556
+ },
2557
+ {
2558
+ "epoch": 4.633339392837666,
2559
+ "grad_norm": 0.8279890418052673,
2560
+ "learning_rate": 5.445092659877153e-05,
2561
+ "loss": 0.5663,
2562
+ "mean_token_accuracy": 0.8239589306712151,
2563
+ "num_tokens": 140363358.0,
2564
+ "step": 6375
2565
+ },
2566
+ {
2567
+ "epoch": 4.651517905835303,
2568
+ "grad_norm": 0.7528174519538879,
2569
+ "learning_rate": 5.427226987838528e-05,
2570
+ "loss": 0.5668,
2571
+ "mean_token_accuracy": 0.8237173920869827,
2572
+ "num_tokens": 140926439.0,
2573
+ "step": 6400
2574
+ },
2575
+ {
2576
+ "epoch": 4.66969641883294,
2577
+ "grad_norm": 0.7415145039558411,
2578
+ "learning_rate": 5.4093286307639786e-05,
2579
+ "loss": 0.559,
2580
+ "mean_token_accuracy": 0.8262969407439232,
2581
+ "num_tokens": 141476080.0,
2582
+ "step": 6425
2583
+ },
2584
+ {
2585
+ "epoch": 4.687874931830576,
2586
+ "grad_norm": 0.7791749238967896,
2587
+ "learning_rate": 5.39139799854517e-05,
2588
+ "loss": 0.5695,
2589
+ "mean_token_accuracy": 0.8218407195806503,
2590
+ "num_tokens": 142035408.0,
2591
+ "step": 6450
2592
+ },
2593
+ {
2594
+ "epoch": 4.706053444828213,
2595
+ "grad_norm": 0.8288747668266296,
2596
+ "learning_rate": 5.3734355018128974e-05,
2597
+ "loss": 0.5658,
2598
+ "mean_token_accuracy": 0.8245827797055244,
2599
+ "num_tokens": 142577709.0,
2600
+ "step": 6475
2601
+ },
2602
+ {
2603
+ "epoch": 4.72423195782585,
2604
+ "grad_norm": 0.7962896227836609,
2605
+ "learning_rate": 5.355441551927695e-05,
2606
+ "loss": 0.5666,
2607
+ "mean_token_accuracy": 0.8251102951169014,
2608
+ "num_tokens": 143123639.0,
2609
+ "step": 6500
2610
+ },
2611
+ {
2612
+ "epoch": 4.72423195782585,
2613
+ "eval_loss": 0.5727137923240662,
2614
+ "eval_mean_token_accuracy": 0.8213643729686737,
2615
+ "eval_num_tokens": 143123639.0,
2616
+ "eval_runtime": 113.5915,
2617
+ "eval_samples_per_second": 43.049,
2618
+ "eval_steps_per_second": 5.388,
2619
+ "step": 6500
2620
  }
2621
  ],
2622
  "logging_steps": 25,
 
2636
  "attributes": {}
2637
  }
2638
  },
2639
+ "total_flos": 3.610250824258683e+17,
2640
  "train_batch_size": 4,
2641
  "trial_name": null,
2642
  "trial_params": null