| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.7472527472527473, |
| "eval_steps": 500, |
| "global_step": 5000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.027472527472527472, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.00019800000000000002, |
| "loss": 1.0178, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.054945054945054944, |
| "grad_norm": 2.796875, |
| "learning_rate": 0.000196, |
| "loss": 1.1549, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.08241758241758242, |
| "grad_norm": 0.00921630859375, |
| "learning_rate": 0.000194, |
| "loss": 0.6677, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.10989010989010989, |
| "grad_norm": 5.5625, |
| "learning_rate": 0.000192, |
| "loss": 0.932, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.13736263736263737, |
| "grad_norm": 3.59375, |
| "learning_rate": 0.00019, |
| "loss": 0.5178, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.16483516483516483, |
| "grad_norm": 5.5625, |
| "learning_rate": 0.000188, |
| "loss": 0.697, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.19230769230769232, |
| "grad_norm": 0.000156402587890625, |
| "learning_rate": 0.00018600000000000002, |
| "loss": 0.7214, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.21978021978021978, |
| "grad_norm": 4.84375, |
| "learning_rate": 0.00018400000000000003, |
| "loss": 0.7486, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.24725274725274726, |
| "grad_norm": 0.447265625, |
| "learning_rate": 0.000182, |
| "loss": 0.6696, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.27472527472527475, |
| "grad_norm": 0.000614166259765625, |
| "learning_rate": 0.00018, |
| "loss": 0.6315, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.3021978021978022, |
| "grad_norm": 8.4375, |
| "learning_rate": 0.00017800000000000002, |
| "loss": 0.652, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.32967032967032966, |
| "grad_norm": 0.000438690185546875, |
| "learning_rate": 0.00017600000000000002, |
| "loss": 0.5758, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.35714285714285715, |
| "grad_norm": 0.0093994140625, |
| "learning_rate": 0.000174, |
| "loss": 0.562, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.38461538461538464, |
| "grad_norm": 0.2109375, |
| "learning_rate": 0.000172, |
| "loss": 0.5937, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.41208791208791207, |
| "grad_norm": 0.0032501220703125, |
| "learning_rate": 0.00017, |
| "loss": 0.6657, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.43956043956043955, |
| "grad_norm": 8.8125, |
| "learning_rate": 0.000168, |
| "loss": 0.7224, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.46703296703296704, |
| "grad_norm": 2.703125, |
| "learning_rate": 0.000166, |
| "loss": 0.9437, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.4945054945054945, |
| "grad_norm": 0.00090789794921875, |
| "learning_rate": 0.000164, |
| "loss": 0.5451, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.521978021978022, |
| "grad_norm": 87.0, |
| "learning_rate": 0.000162, |
| "loss": 0.4207, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.5494505494505495, |
| "grad_norm": 0.00164031982421875, |
| "learning_rate": 0.00016, |
| "loss": 0.5841, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.5769230769230769, |
| "grad_norm": 0.00034332275390625, |
| "learning_rate": 0.00015800000000000002, |
| "loss": 0.5337, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.6043956043956044, |
| "grad_norm": 0.1201171875, |
| "learning_rate": 0.00015600000000000002, |
| "loss": 0.7521, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.6318681318681318, |
| "grad_norm": 0.05078125, |
| "learning_rate": 0.000154, |
| "loss": 0.6175, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.6593406593406593, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.000152, |
| "loss": 0.5597, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.6868131868131868, |
| "grad_norm": 0.06591796875, |
| "learning_rate": 0.00015000000000000001, |
| "loss": 0.9526, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.7142857142857143, |
| "grad_norm": 2.703125, |
| "learning_rate": 0.000148, |
| "loss": 0.416, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.7417582417582418, |
| "grad_norm": 0.01177978515625, |
| "learning_rate": 0.000146, |
| "loss": 0.6454, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.7692307692307693, |
| "grad_norm": 0.423828125, |
| "learning_rate": 0.000144, |
| "loss": 0.5772, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.7967032967032966, |
| "grad_norm": 2.171875, |
| "learning_rate": 0.000142, |
| "loss": 0.5839, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.8241758241758241, |
| "grad_norm": 0.003021240234375, |
| "learning_rate": 0.00014, |
| "loss": 0.7548, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.8516483516483516, |
| "grad_norm": 4.6875, |
| "learning_rate": 0.000138, |
| "loss": 0.678, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.8791208791208791, |
| "grad_norm": 0.00433349609375, |
| "learning_rate": 0.00013600000000000003, |
| "loss": 0.5795, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.9065934065934066, |
| "grad_norm": 0.00156402587890625, |
| "learning_rate": 0.000134, |
| "loss": 0.6352, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.9340659340659341, |
| "grad_norm": 6.875, |
| "learning_rate": 0.000132, |
| "loss": 0.5606, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.9615384615384616, |
| "grad_norm": 3.46875, |
| "learning_rate": 0.00013000000000000002, |
| "loss": 0.5033, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.989010989010989, |
| "grad_norm": 7.53125, |
| "learning_rate": 0.00012800000000000002, |
| "loss": 0.3787, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.0164835164835164, |
| "grad_norm": 8.625, |
| "learning_rate": 0.000126, |
| "loss": 0.5048, |
| "step": 1850 |
| }, |
| { |
| "epoch": 1.043956043956044, |
| "grad_norm": 0.000518798828125, |
| "learning_rate": 0.000124, |
| "loss": 0.5134, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.0714285714285714, |
| "grad_norm": 0.008056640625, |
| "learning_rate": 0.000122, |
| "loss": 0.2342, |
| "step": 1950 |
| }, |
| { |
| "epoch": 1.098901098901099, |
| "grad_norm": 4.78125, |
| "learning_rate": 0.00012, |
| "loss": 0.4388, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.1263736263736264, |
| "grad_norm": 7.343292236328125e-05, |
| "learning_rate": 0.000118, |
| "loss": 0.4928, |
| "step": 2050 |
| }, |
| { |
| "epoch": 1.1538461538461537, |
| "grad_norm": 7.34375, |
| "learning_rate": 0.000116, |
| "loss": 0.4759, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.1813186813186813, |
| "grad_norm": 5.3125, |
| "learning_rate": 0.00011399999999999999, |
| "loss": 0.3002, |
| "step": 2150 |
| }, |
| { |
| "epoch": 1.2087912087912087, |
| "grad_norm": 0.00025177001953125, |
| "learning_rate": 0.00011200000000000001, |
| "loss": 0.1603, |
| "step": 2200 |
| }, |
| { |
| "epoch": 1.2362637362637363, |
| "grad_norm": 0.0003719329833984375, |
| "learning_rate": 0.00011000000000000002, |
| "loss": 0.3096, |
| "step": 2250 |
| }, |
| { |
| "epoch": 1.2637362637362637, |
| "grad_norm": 0.000732421875, |
| "learning_rate": 0.00010800000000000001, |
| "loss": 0.2216, |
| "step": 2300 |
| }, |
| { |
| "epoch": 1.2912087912087913, |
| "grad_norm": 7.033348083496094e-06, |
| "learning_rate": 0.00010600000000000002, |
| "loss": 0.3195, |
| "step": 2350 |
| }, |
| { |
| "epoch": 1.3186813186813187, |
| "grad_norm": 2.0384788513183594e-05, |
| "learning_rate": 0.00010400000000000001, |
| "loss": 0.2727, |
| "step": 2400 |
| }, |
| { |
| "epoch": 1.3461538461538463, |
| "grad_norm": 0.00054931640625, |
| "learning_rate": 0.00010200000000000001, |
| "loss": 0.3244, |
| "step": 2450 |
| }, |
| { |
| "epoch": 1.3736263736263736, |
| "grad_norm": 10.9375, |
| "learning_rate": 0.0001, |
| "loss": 0.2455, |
| "step": 2500 |
| }, |
| { |
| "epoch": 1.401098901098901, |
| "grad_norm": 8.535385131835938e-05, |
| "learning_rate": 9.8e-05, |
| "loss": 0.3644, |
| "step": 2550 |
| }, |
| { |
| "epoch": 1.4285714285714286, |
| "grad_norm": 0.00012969970703125, |
| "learning_rate": 9.6e-05, |
| "loss": 0.3234, |
| "step": 2600 |
| }, |
| { |
| "epoch": 1.456043956043956, |
| "grad_norm": 6.389617919921875e-05, |
| "learning_rate": 9.4e-05, |
| "loss": 0.1711, |
| "step": 2650 |
| }, |
| { |
| "epoch": 1.4835164835164836, |
| "grad_norm": 7.96875, |
| "learning_rate": 9.200000000000001e-05, |
| "loss": 0.3783, |
| "step": 2700 |
| }, |
| { |
| "epoch": 1.510989010989011, |
| "grad_norm": 15.25, |
| "learning_rate": 9e-05, |
| "loss": 0.2034, |
| "step": 2750 |
| }, |
| { |
| "epoch": 1.5384615384615383, |
| "grad_norm": 4.96875, |
| "learning_rate": 8.800000000000001e-05, |
| "loss": 0.2156, |
| "step": 2800 |
| }, |
| { |
| "epoch": 1.565934065934066, |
| "grad_norm": 9.107589721679688e-05, |
| "learning_rate": 8.6e-05, |
| "loss": 0.3119, |
| "step": 2850 |
| }, |
| { |
| "epoch": 1.5934065934065935, |
| "grad_norm": 0.0002079010009765625, |
| "learning_rate": 8.4e-05, |
| "loss": 0.2044, |
| "step": 2900 |
| }, |
| { |
| "epoch": 1.620879120879121, |
| "grad_norm": 0.11376953125, |
| "learning_rate": 8.2e-05, |
| "loss": 0.2696, |
| "step": 2950 |
| }, |
| { |
| "epoch": 1.6483516483516483, |
| "grad_norm": 10.375, |
| "learning_rate": 8e-05, |
| "loss": 0.2998, |
| "step": 3000 |
| }, |
| { |
| "epoch": 1.6758241758241759, |
| "grad_norm": 0.00054931640625, |
| "learning_rate": 7.800000000000001e-05, |
| "loss": 0.2754, |
| "step": 3050 |
| }, |
| { |
| "epoch": 1.7032967032967035, |
| "grad_norm": 8.0108642578125e-05, |
| "learning_rate": 7.6e-05, |
| "loss": 0.2065, |
| "step": 3100 |
| }, |
| { |
| "epoch": 1.7307692307692308, |
| "grad_norm": 0.00016307830810546875, |
| "learning_rate": 7.4e-05, |
| "loss": 0.2248, |
| "step": 3150 |
| }, |
| { |
| "epoch": 1.7582417582417582, |
| "grad_norm": 0.00012969970703125, |
| "learning_rate": 7.2e-05, |
| "loss": 0.3116, |
| "step": 3200 |
| }, |
| { |
| "epoch": 1.7857142857142856, |
| "grad_norm": 0.0002231597900390625, |
| "learning_rate": 7e-05, |
| "loss": 0.2603, |
| "step": 3250 |
| }, |
| { |
| "epoch": 1.8131868131868132, |
| "grad_norm": 0.00012874603271484375, |
| "learning_rate": 6.800000000000001e-05, |
| "loss": 0.2854, |
| "step": 3300 |
| }, |
| { |
| "epoch": 1.8406593406593408, |
| "grad_norm": 3.96875, |
| "learning_rate": 6.6e-05, |
| "loss": 0.2222, |
| "step": 3350 |
| }, |
| { |
| "epoch": 1.8681318681318682, |
| "grad_norm": 0.0033416748046875, |
| "learning_rate": 6.400000000000001e-05, |
| "loss": 0.2902, |
| "step": 3400 |
| }, |
| { |
| "epoch": 1.8956043956043955, |
| "grad_norm": 6.008148193359375e-05, |
| "learning_rate": 6.2e-05, |
| "loss": 0.2299, |
| "step": 3450 |
| }, |
| { |
| "epoch": 1.9230769230769231, |
| "grad_norm": 18.75, |
| "learning_rate": 6e-05, |
| "loss": 0.2612, |
| "step": 3500 |
| }, |
| { |
| "epoch": 1.9505494505494505, |
| "grad_norm": 6.90625, |
| "learning_rate": 5.8e-05, |
| "loss": 0.2367, |
| "step": 3550 |
| }, |
| { |
| "epoch": 1.978021978021978, |
| "grad_norm": 5.650520324707031e-05, |
| "learning_rate": 5.6000000000000006e-05, |
| "loss": 0.2223, |
| "step": 3600 |
| }, |
| { |
| "epoch": 2.0054945054945055, |
| "grad_norm": 0.03564453125, |
| "learning_rate": 5.4000000000000005e-05, |
| "loss": 0.2567, |
| "step": 3650 |
| }, |
| { |
| "epoch": 2.032967032967033, |
| "grad_norm": 0.0002574920654296875, |
| "learning_rate": 5.2000000000000004e-05, |
| "loss": 0.1277, |
| "step": 3700 |
| }, |
| { |
| "epoch": 2.0604395604395602, |
| "grad_norm": 0.00070953369140625, |
| "learning_rate": 5e-05, |
| "loss": 0.1751, |
| "step": 3750 |
| }, |
| { |
| "epoch": 2.087912087912088, |
| "grad_norm": 0.00015735626220703125, |
| "learning_rate": 4.8e-05, |
| "loss": 0.1713, |
| "step": 3800 |
| }, |
| { |
| "epoch": 2.1153846153846154, |
| "grad_norm": 0.00021648406982421875, |
| "learning_rate": 4.600000000000001e-05, |
| "loss": 0.1682, |
| "step": 3850 |
| }, |
| { |
| "epoch": 2.142857142857143, |
| "grad_norm": 4.380941390991211e-06, |
| "learning_rate": 4.4000000000000006e-05, |
| "loss": 0.1259, |
| "step": 3900 |
| }, |
| { |
| "epoch": 2.17032967032967, |
| "grad_norm": 8.3125, |
| "learning_rate": 4.2e-05, |
| "loss": 0.1822, |
| "step": 3950 |
| }, |
| { |
| "epoch": 2.197802197802198, |
| "grad_norm": 0.0003376007080078125, |
| "learning_rate": 4e-05, |
| "loss": 0.105, |
| "step": 4000 |
| }, |
| { |
| "epoch": 2.2252747252747254, |
| "grad_norm": 8.726119995117188e-05, |
| "learning_rate": 3.8e-05, |
| "loss": 0.0887, |
| "step": 4050 |
| }, |
| { |
| "epoch": 2.2527472527472527, |
| "grad_norm": 0.00012493133544921875, |
| "learning_rate": 3.6e-05, |
| "loss": 0.1371, |
| "step": 4100 |
| }, |
| { |
| "epoch": 2.28021978021978, |
| "grad_norm": 3.25, |
| "learning_rate": 3.4000000000000007e-05, |
| "loss": 0.1725, |
| "step": 4150 |
| }, |
| { |
| "epoch": 2.3076923076923075, |
| "grad_norm": 0.00016117095947265625, |
| "learning_rate": 3.2000000000000005e-05, |
| "loss": 0.1363, |
| "step": 4200 |
| }, |
| { |
| "epoch": 2.3351648351648353, |
| "grad_norm": 4.125, |
| "learning_rate": 3e-05, |
| "loss": 0.203, |
| "step": 4250 |
| }, |
| { |
| "epoch": 2.3626373626373627, |
| "grad_norm": 3.796875, |
| "learning_rate": 2.8000000000000003e-05, |
| "loss": 0.1607, |
| "step": 4300 |
| }, |
| { |
| "epoch": 2.39010989010989, |
| "grad_norm": 2.8908252716064453e-06, |
| "learning_rate": 2.6000000000000002e-05, |
| "loss": 0.0885, |
| "step": 4350 |
| }, |
| { |
| "epoch": 2.4175824175824174, |
| "grad_norm": 2.046875, |
| "learning_rate": 2.4e-05, |
| "loss": 0.1202, |
| "step": 4400 |
| }, |
| { |
| "epoch": 2.4450549450549453, |
| "grad_norm": 1.0132789611816406e-05, |
| "learning_rate": 2.2000000000000003e-05, |
| "loss": 0.2216, |
| "step": 4450 |
| }, |
| { |
| "epoch": 2.4725274725274726, |
| "grad_norm": 2.265625, |
| "learning_rate": 2e-05, |
| "loss": 0.1367, |
| "step": 4500 |
| }, |
| { |
| "epoch": 2.5, |
| "grad_norm": 2.5, |
| "learning_rate": 1.8e-05, |
| "loss": 0.1546, |
| "step": 4550 |
| }, |
| { |
| "epoch": 2.5274725274725274, |
| "grad_norm": 0.00122833251953125, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 0.1059, |
| "step": 4600 |
| }, |
| { |
| "epoch": 2.5549450549450547, |
| "grad_norm": 0.0004730224609375, |
| "learning_rate": 1.4000000000000001e-05, |
| "loss": 0.0869, |
| "step": 4650 |
| }, |
| { |
| "epoch": 2.5824175824175826, |
| "grad_norm": 9.625, |
| "learning_rate": 1.2e-05, |
| "loss": 0.1142, |
| "step": 4700 |
| }, |
| { |
| "epoch": 2.60989010989011, |
| "grad_norm": 4.15625, |
| "learning_rate": 1e-05, |
| "loss": 0.2082, |
| "step": 4750 |
| }, |
| { |
| "epoch": 2.6373626373626373, |
| "grad_norm": 5.03125, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 0.1374, |
| "step": 4800 |
| }, |
| { |
| "epoch": 2.6648351648351647, |
| "grad_norm": 2.46875, |
| "learning_rate": 6e-06, |
| "loss": 0.1475, |
| "step": 4850 |
| }, |
| { |
| "epoch": 2.6923076923076925, |
| "grad_norm": 3.125, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 0.1593, |
| "step": 4900 |
| }, |
| { |
| "epoch": 2.71978021978022, |
| "grad_norm": 7.295608520507812e-05, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 0.1695, |
| "step": 4950 |
| }, |
| { |
| "epoch": 2.7472527472527473, |
| "grad_norm": 6.6875, |
| "learning_rate": 0.0, |
| "loss": 0.1467, |
| "step": 5000 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 5000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.36023173482496e+16, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|