| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 250, |
| "global_step": 1000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.001, |
| "grad_norm": 55.0, |
| "learning_rate": 0.0, |
| "loss": 15.8026, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.002, |
| "grad_norm": 53.5, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 15.7515, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.003, |
| "grad_norm": 54.0, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 15.6805, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.004, |
| "grad_norm": 53.75, |
| "learning_rate": 6e-06, |
| "loss": 15.6899, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.005, |
| "grad_norm": 51.75, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 15.4635, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.006, |
| "grad_norm": 52.0, |
| "learning_rate": 1e-05, |
| "loss": 15.2861, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.007, |
| "grad_norm": 50.25, |
| "learning_rate": 1.2e-05, |
| "loss": 15.0195, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.008, |
| "grad_norm": 51.75, |
| "learning_rate": 1.4000000000000001e-05, |
| "loss": 14.5896, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.009, |
| "grad_norm": 48.5, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 14.428, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.01, |
| "grad_norm": 43.0, |
| "learning_rate": 1.8e-05, |
| "loss": 13.8742, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.011, |
| "grad_norm": 35.75, |
| "learning_rate": 2e-05, |
| "loss": 13.1404, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.012, |
| "grad_norm": 25.75, |
| "learning_rate": 2.2000000000000003e-05, |
| "loss": 12.7393, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.013, |
| "grad_norm": 28.25, |
| "learning_rate": 2.4e-05, |
| "loss": 12.4467, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.014, |
| "grad_norm": 39.75, |
| "learning_rate": 2.6000000000000002e-05, |
| "loss": 12.3137, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.015, |
| "grad_norm": 40.75, |
| "learning_rate": 2.8000000000000003e-05, |
| "loss": 11.8957, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.016, |
| "grad_norm": 39.75, |
| "learning_rate": 3e-05, |
| "loss": 11.3899, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.017, |
| "grad_norm": 32.0, |
| "learning_rate": 3.2000000000000005e-05, |
| "loss": 10.4023, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.018, |
| "grad_norm": 25.0, |
| "learning_rate": 3.4000000000000007e-05, |
| "loss": 9.4691, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.019, |
| "grad_norm": 23.5, |
| "learning_rate": 3.6e-05, |
| "loss": 8.8468, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.02, |
| "grad_norm": 20.5, |
| "learning_rate": 3.8e-05, |
| "loss": 8.201, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.021, |
| "grad_norm": 17.5, |
| "learning_rate": 4e-05, |
| "loss": 7.7326, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.022, |
| "grad_norm": 17.375, |
| "learning_rate": 4.2e-05, |
| "loss": 7.3021, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.023, |
| "grad_norm": 14.8125, |
| "learning_rate": 4.4000000000000006e-05, |
| "loss": 7.0761, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.024, |
| "grad_norm": 12.0625, |
| "learning_rate": 4.600000000000001e-05, |
| "loss": 7.0736, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.025, |
| "grad_norm": 9.625, |
| "learning_rate": 4.8e-05, |
| "loss": 6.6617, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.026, |
| "grad_norm": 7.125, |
| "learning_rate": 5e-05, |
| "loss": 6.2627, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.027, |
| "grad_norm": 6.0, |
| "learning_rate": 5.2000000000000004e-05, |
| "loss": 6.108, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.028, |
| "grad_norm": 4.65625, |
| "learning_rate": 5.4000000000000005e-05, |
| "loss": 5.7079, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.029, |
| "grad_norm": 4.5625, |
| "learning_rate": 5.6000000000000006e-05, |
| "loss": 6.0111, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.03, |
| "grad_norm": 3.78125, |
| "learning_rate": 5.8e-05, |
| "loss": 6.0342, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.031, |
| "grad_norm": 3.4375, |
| "learning_rate": 6e-05, |
| "loss": 5.8238, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.032, |
| "grad_norm": 3.4375, |
| "learning_rate": 6.2e-05, |
| "loss": 5.732, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.033, |
| "grad_norm": 3.171875, |
| "learning_rate": 6.400000000000001e-05, |
| "loss": 5.3643, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.034, |
| "grad_norm": 3.546875, |
| "learning_rate": 6.6e-05, |
| "loss": 5.2917, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.035, |
| "grad_norm": 4.21875, |
| "learning_rate": 6.800000000000001e-05, |
| "loss": 5.2339, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.036, |
| "grad_norm": 2.859375, |
| "learning_rate": 7e-05, |
| "loss": 5.2663, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.037, |
| "grad_norm": 2.125, |
| "learning_rate": 7.2e-05, |
| "loss": 5.0012, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.038, |
| "grad_norm": 2.84375, |
| "learning_rate": 7.4e-05, |
| "loss": 4.8893, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.039, |
| "grad_norm": 4.09375, |
| "learning_rate": 7.6e-05, |
| "loss": 4.9856, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.04, |
| "grad_norm": 2.046875, |
| "learning_rate": 7.800000000000001e-05, |
| "loss": 5.0183, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.041, |
| "grad_norm": 8.875, |
| "learning_rate": 8e-05, |
| "loss": 4.9491, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.042, |
| "grad_norm": 4.65625, |
| "learning_rate": 8.2e-05, |
| "loss": 4.7829, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.043, |
| "grad_norm": 4.40625, |
| "learning_rate": 8.4e-05, |
| "loss": 4.9717, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.044, |
| "grad_norm": 4.34375, |
| "learning_rate": 8.6e-05, |
| "loss": 4.817, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.045, |
| "grad_norm": 2.1875, |
| "learning_rate": 8.800000000000001e-05, |
| "loss": 4.8171, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.046, |
| "grad_norm": 5.1875, |
| "learning_rate": 9e-05, |
| "loss": 4.8199, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.047, |
| "grad_norm": 3.40625, |
| "learning_rate": 9.200000000000001e-05, |
| "loss": 4.7965, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.048, |
| "grad_norm": 3.609375, |
| "learning_rate": 9.4e-05, |
| "loss": 4.649, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.049, |
| "grad_norm": 3.59375, |
| "learning_rate": 9.6e-05, |
| "loss": 4.6714, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.05, |
| "grad_norm": 3.25, |
| "learning_rate": 9.8e-05, |
| "loss": 4.5177, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.051, |
| "grad_norm": 3.03125, |
| "learning_rate": 0.0001, |
| "loss": 4.5112, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.052, |
| "grad_norm": 2.15625, |
| "learning_rate": 0.00010200000000000001, |
| "loss": 4.5739, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.053, |
| "grad_norm": 2.1875, |
| "learning_rate": 0.00010400000000000001, |
| "loss": 4.4601, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.054, |
| "grad_norm": 2.46875, |
| "learning_rate": 0.00010600000000000002, |
| "loss": 4.5269, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.055, |
| "grad_norm": 1.8828125, |
| "learning_rate": 0.00010800000000000001, |
| "loss": 4.2753, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.056, |
| "grad_norm": 3.203125, |
| "learning_rate": 0.00011000000000000002, |
| "loss": 4.3292, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.057, |
| "grad_norm": 2.390625, |
| "learning_rate": 0.00011200000000000001, |
| "loss": 4.3489, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.058, |
| "grad_norm": 4.125, |
| "learning_rate": 0.00011399999999999999, |
| "loss": 4.4857, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.059, |
| "grad_norm": 2.96875, |
| "learning_rate": 0.000116, |
| "loss": 4.3047, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.06, |
| "grad_norm": 3.265625, |
| "learning_rate": 0.000118, |
| "loss": 4.4213, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.061, |
| "grad_norm": 11.4375, |
| "learning_rate": 0.00012, |
| "loss": 4.3308, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.062, |
| "grad_norm": 2.59375, |
| "learning_rate": 0.000122, |
| "loss": 4.3524, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.063, |
| "grad_norm": 3.375, |
| "learning_rate": 0.000124, |
| "loss": 4.2236, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.064, |
| "grad_norm": 2.234375, |
| "learning_rate": 0.000126, |
| "loss": 4.3674, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.065, |
| "grad_norm": 3.0, |
| "learning_rate": 0.00012800000000000002, |
| "loss": 4.1765, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.066, |
| "grad_norm": 2.234375, |
| "learning_rate": 0.00013000000000000002, |
| "loss": 4.1043, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.067, |
| "grad_norm": 3.8125, |
| "learning_rate": 0.000132, |
| "loss": 4.0828, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.068, |
| "grad_norm": 2.75, |
| "learning_rate": 0.000134, |
| "loss": 4.1654, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.069, |
| "grad_norm": 3.109375, |
| "learning_rate": 0.00013600000000000003, |
| "loss": 4.0984, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.07, |
| "grad_norm": 2.171875, |
| "learning_rate": 0.000138, |
| "loss": 4.0978, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.071, |
| "grad_norm": 3.84375, |
| "learning_rate": 0.00014, |
| "loss": 4.1044, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.072, |
| "grad_norm": 2.5625, |
| "learning_rate": 0.000142, |
| "loss": 3.967, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.073, |
| "grad_norm": 2.0625, |
| "learning_rate": 0.000144, |
| "loss": 4.0976, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.074, |
| "grad_norm": 2.328125, |
| "learning_rate": 0.000146, |
| "loss": 3.8734, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.075, |
| "grad_norm": 2.015625, |
| "learning_rate": 0.000148, |
| "loss": 3.9381, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.076, |
| "grad_norm": 2.921875, |
| "learning_rate": 0.00015000000000000001, |
| "loss": 3.97, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.077, |
| "grad_norm": 2.046875, |
| "learning_rate": 0.000152, |
| "loss": 3.9806, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.078, |
| "grad_norm": 3.25, |
| "learning_rate": 0.000154, |
| "loss": 3.7804, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.079, |
| "grad_norm": 2.578125, |
| "learning_rate": 0.00015600000000000002, |
| "loss": 3.9961, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 1.9765625, |
| "learning_rate": 0.00015800000000000002, |
| "loss": 3.9899, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.081, |
| "grad_norm": 2.171875, |
| "learning_rate": 0.00016, |
| "loss": 3.8394, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.082, |
| "grad_norm": 1.8671875, |
| "learning_rate": 0.000162, |
| "loss": 3.8614, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.083, |
| "grad_norm": 2.140625, |
| "learning_rate": 0.000164, |
| "loss": 3.7687, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.084, |
| "grad_norm": 1.5546875, |
| "learning_rate": 0.000166, |
| "loss": 3.8413, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.085, |
| "grad_norm": 2.375, |
| "learning_rate": 0.000168, |
| "loss": 3.7533, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.086, |
| "grad_norm": 1.890625, |
| "learning_rate": 0.00017, |
| "loss": 3.8058, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.087, |
| "grad_norm": 1.90625, |
| "learning_rate": 0.000172, |
| "loss": 3.8243, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.088, |
| "grad_norm": 2.25, |
| "learning_rate": 0.000174, |
| "loss": 3.7795, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.089, |
| "grad_norm": 1.8359375, |
| "learning_rate": 0.00017600000000000002, |
| "loss": 3.7421, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.09, |
| "grad_norm": 1.7421875, |
| "learning_rate": 0.00017800000000000002, |
| "loss": 3.7222, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.091, |
| "grad_norm": 1.5078125, |
| "learning_rate": 0.00018, |
| "loss": 3.727, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.092, |
| "grad_norm": 2.328125, |
| "learning_rate": 0.000182, |
| "loss": 3.8749, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.093, |
| "grad_norm": 1.7734375, |
| "learning_rate": 0.00018400000000000003, |
| "loss": 3.6911, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.094, |
| "grad_norm": 2.4375, |
| "learning_rate": 0.00018600000000000002, |
| "loss": 3.6221, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.095, |
| "grad_norm": 2.265625, |
| "learning_rate": 0.000188, |
| "loss": 3.6225, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.096, |
| "grad_norm": 2.078125, |
| "learning_rate": 0.00019, |
| "loss": 3.5456, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.097, |
| "grad_norm": 3.0625, |
| "learning_rate": 0.000192, |
| "loss": 3.5316, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.098, |
| "grad_norm": 1.75, |
| "learning_rate": 0.000194, |
| "loss": 3.5942, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.099, |
| "grad_norm": 1.5078125, |
| "learning_rate": 0.000196, |
| "loss": 3.5513, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.1, |
| "grad_norm": 2.8125, |
| "learning_rate": 0.00019800000000000002, |
| "loss": 3.673, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.101, |
| "grad_norm": 2.328125, |
| "learning_rate": 0.0002, |
| "loss": 3.6088, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.102, |
| "grad_norm": 1.4921875, |
| "learning_rate": 0.00019999939076577905, |
| "loss": 3.5971, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.103, |
| "grad_norm": 1.421875, |
| "learning_rate": 0.00019999756307053948, |
| "loss": 3.5318, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.104, |
| "grad_norm": 2.046875, |
| "learning_rate": 0.00019999451693655123, |
| "loss": 3.5199, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.105, |
| "grad_norm": 1.7734375, |
| "learning_rate": 0.00019999025240093044, |
| "loss": 3.5049, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.106, |
| "grad_norm": 2.234375, |
| "learning_rate": 0.00019998476951563915, |
| "loss": 3.5481, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.107, |
| "grad_norm": 1.9921875, |
| "learning_rate": 0.00019997806834748456, |
| "loss": 3.4391, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.108, |
| "grad_norm": 1.4609375, |
| "learning_rate": 0.00019997014897811833, |
| "loss": 3.5096, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.109, |
| "grad_norm": 1.8046875, |
| "learning_rate": 0.00019996101150403543, |
| "loss": 3.424, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.11, |
| "grad_norm": 2.4375, |
| "learning_rate": 0.00019995065603657316, |
| "loss": 3.4862, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.111, |
| "grad_norm": 1.5859375, |
| "learning_rate": 0.0001999390827019096, |
| "loss": 3.5117, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.112, |
| "grad_norm": 1.7265625, |
| "learning_rate": 0.0001999262916410621, |
| "loss": 3.452, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.113, |
| "grad_norm": 1.7578125, |
| "learning_rate": 0.00019991228300988585, |
| "loss": 3.4834, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.114, |
| "grad_norm": 1.8984375, |
| "learning_rate": 0.00019989705697907149, |
| "loss": 3.4673, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.115, |
| "grad_norm": 1.625, |
| "learning_rate": 0.0001998806137341434, |
| "loss": 3.2031, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.116, |
| "grad_norm": 1.6015625, |
| "learning_rate": 0.0001998629534754574, |
| "loss": 3.3329, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.117, |
| "grad_norm": 1.4921875, |
| "learning_rate": 0.00019984407641819812, |
| "loss": 3.4215, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.118, |
| "grad_norm": 1.65625, |
| "learning_rate": 0.00019982398279237655, |
| "loss": 3.3886, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.119, |
| "grad_norm": 1.421875, |
| "learning_rate": 0.00019980267284282717, |
| "loss": 3.2904, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.12, |
| "grad_norm": 1.3515625, |
| "learning_rate": 0.000199780146829205, |
| "loss": 3.3135, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.121, |
| "grad_norm": 1.8671875, |
| "learning_rate": 0.00019975640502598244, |
| "loss": 3.4523, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.122, |
| "grad_norm": 1.796875, |
| "learning_rate": 0.00019973144772244582, |
| "loss": 3.248, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.123, |
| "grad_norm": 1.4140625, |
| "learning_rate": 0.00019970527522269205, |
| "loss": 3.3419, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.124, |
| "grad_norm": 1.453125, |
| "learning_rate": 0.00019967788784562473, |
| "loss": 3.296, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.125, |
| "grad_norm": 1.90625, |
| "learning_rate": 0.00019964928592495045, |
| "loss": 3.2151, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.126, |
| "grad_norm": 1.3828125, |
| "learning_rate": 0.00019961946980917456, |
| "loss": 3.3786, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.127, |
| "grad_norm": 1.859375, |
| "learning_rate": 0.00019958843986159704, |
| "loss": 3.4188, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.128, |
| "grad_norm": 1.46875, |
| "learning_rate": 0.00019955619646030802, |
| "loss": 3.4489, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.129, |
| "grad_norm": 1.7265625, |
| "learning_rate": 0.0001995227399981831, |
| "loss": 3.3746, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.13, |
| "grad_norm": 1.921875, |
| "learning_rate": 0.00019948807088287883, |
| "loss": 3.29, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.131, |
| "grad_norm": 1.3984375, |
| "learning_rate": 0.00019945218953682734, |
| "loss": 3.2717, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.132, |
| "grad_norm": 4.21875, |
| "learning_rate": 0.00019941509639723155, |
| "loss": 3.2601, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.133, |
| "grad_norm": 2.15625, |
| "learning_rate": 0.00019937679191605963, |
| "loss": 3.3615, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.134, |
| "grad_norm": 2.390625, |
| "learning_rate": 0.00019933727656003963, |
| "loss": 3.4326, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.135, |
| "grad_norm": 1.9296875, |
| "learning_rate": 0.0001992965508106537, |
| "loss": 3.3733, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.136, |
| "grad_norm": 1.6953125, |
| "learning_rate": 0.00019925461516413223, |
| "loss": 3.3488, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.137, |
| "grad_norm": 1.46875, |
| "learning_rate": 0.0001992114701314478, |
| "loss": 3.2991, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.138, |
| "grad_norm": 1.375, |
| "learning_rate": 0.00019916711623830903, |
| "loss": 3.2037, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.139, |
| "grad_norm": 1.5234375, |
| "learning_rate": 0.00019912155402515417, |
| "loss": 3.2748, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.14, |
| "grad_norm": 1.4375, |
| "learning_rate": 0.00019907478404714436, |
| "loss": 3.2753, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.141, |
| "grad_norm": 2.0, |
| "learning_rate": 0.00019902680687415705, |
| "loss": 3.3406, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.142, |
| "grad_norm": 1.4296875, |
| "learning_rate": 0.0001989776230907789, |
| "loss": 3.3422, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.143, |
| "grad_norm": 1.703125, |
| "learning_rate": 0.00019892723329629887, |
| "loss": 3.4146, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.144, |
| "grad_norm": 1.4453125, |
| "learning_rate": 0.0001988756381047006, |
| "loss": 3.3321, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.145, |
| "grad_norm": 1.46875, |
| "learning_rate": 0.0001988228381446553, |
| "loss": 3.3225, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.146, |
| "grad_norm": 1.3984375, |
| "learning_rate": 0.00019876883405951377, |
| "loss": 3.2079, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.147, |
| "grad_norm": 1.59375, |
| "learning_rate": 0.0001987136265072988, |
| "loss": 3.1757, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.148, |
| "grad_norm": 1.40625, |
| "learning_rate": 0.00019865721616069696, |
| "loss": 3.1882, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.149, |
| "grad_norm": 1.3671875, |
| "learning_rate": 0.0001985996037070505, |
| "loss": 3.2314, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.15, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00019854078984834903, |
| "loss": 3.1942, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.151, |
| "grad_norm": 1.4453125, |
| "learning_rate": 0.00019848077530122083, |
| "loss": 3.2477, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.152, |
| "grad_norm": 1.7734375, |
| "learning_rate": 0.0001984195607969242, |
| "loss": 3.1375, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.153, |
| "grad_norm": 1.484375, |
| "learning_rate": 0.00019835714708133862, |
| "loss": 3.145, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.154, |
| "grad_norm": 1.2265625, |
| "learning_rate": 0.00019829353491495545, |
| "loss": 3.1747, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.155, |
| "grad_norm": 1.3046875, |
| "learning_rate": 0.0001982287250728689, |
| "loss": 3.2027, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.156, |
| "grad_norm": 1.4609375, |
| "learning_rate": 0.00019816271834476642, |
| "loss": 3.09, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.157, |
| "grad_norm": 2.578125, |
| "learning_rate": 0.00019809551553491916, |
| "loss": 3.2477, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.158, |
| "grad_norm": 1.7421875, |
| "learning_rate": 0.00019802711746217218, |
| "loss": 3.1619, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.159, |
| "grad_norm": 1.3984375, |
| "learning_rate": 0.0001979575249599344, |
| "loss": 3.2682, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 1.6328125, |
| "learning_rate": 0.0001978867388761685, |
| "loss": 3.0885, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.161, |
| "grad_norm": 1.2890625, |
| "learning_rate": 0.00019781476007338058, |
| "loss": 3.2491, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.162, |
| "grad_norm": 1.5, |
| "learning_rate": 0.0001977415894286096, |
| "loss": 3.21, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.163, |
| "grad_norm": 1.2734375, |
| "learning_rate": 0.0001976672278334168, |
| "loss": 3.1559, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.164, |
| "grad_norm": 1.3515625, |
| "learning_rate": 0.00019759167619387476, |
| "loss": 3.0837, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.165, |
| "grad_norm": 1.3046875, |
| "learning_rate": 0.00019751493543055632, |
| "loss": 3.1893, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.166, |
| "grad_norm": 1.203125, |
| "learning_rate": 0.00019743700647852354, |
| "loss": 3.1023, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.167, |
| "grad_norm": 1.1953125, |
| "learning_rate": 0.00019735789028731604, |
| "loss": 3.0936, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.168, |
| "grad_norm": 1.359375, |
| "learning_rate": 0.00019727758782093967, |
| "loss": 3.2097, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.169, |
| "grad_norm": 1.6171875, |
| "learning_rate": 0.00019719610005785465, |
| "loss": 3.1198, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.17, |
| "grad_norm": 1.3359375, |
| "learning_rate": 0.00019711342799096361, |
| "loss": 3.1823, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.171, |
| "grad_norm": 1.0703125, |
| "learning_rate": 0.00019702957262759965, |
| "loss": 3.1136, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.172, |
| "grad_norm": 4.90625, |
| "learning_rate": 0.0001969445349895139, |
| "loss": 3.1076, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.173, |
| "grad_norm": 1.625, |
| "learning_rate": 0.0001968583161128631, |
| "loss": 3.113, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.174, |
| "grad_norm": 1.6640625, |
| "learning_rate": 0.00019677091704819715, |
| "loss": 3.1752, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.175, |
| "grad_norm": 1.5390625, |
| "learning_rate": 0.00019668233886044597, |
| "loss": 3.1309, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.176, |
| "grad_norm": 1.34375, |
| "learning_rate": 0.00019659258262890683, |
| "loss": 3.1505, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.177, |
| "grad_norm": 1.3125, |
| "learning_rate": 0.00019650164944723115, |
| "loss": 3.1151, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.178, |
| "grad_norm": 1.5546875, |
| "learning_rate": 0.00019640954042341103, |
| "loss": 3.085, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.179, |
| "grad_norm": 1.421875, |
| "learning_rate": 0.00019631625667976583, |
| "loss": 3.0857, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.18, |
| "grad_norm": 1.1953125, |
| "learning_rate": 0.00019622179935292855, |
| "loss": 3.2549, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.181, |
| "grad_norm": 1.265625, |
| "learning_rate": 0.0001961261695938319, |
| "loss": 3.2138, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.182, |
| "grad_norm": 1.1796875, |
| "learning_rate": 0.0001960293685676943, |
| "loss": 3.0507, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.183, |
| "grad_norm": 1.4453125, |
| "learning_rate": 0.00019593139745400576, |
| "loss": 3.2578, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.184, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00019583225744651333, |
| "loss": 3.0951, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.185, |
| "grad_norm": 1.6875, |
| "learning_rate": 0.00019573194975320673, |
| "loss": 3.1273, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.186, |
| "grad_norm": 1.296875, |
| "learning_rate": 0.00019563047559630357, |
| "loss": 3.0946, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.187, |
| "grad_norm": 1.21875, |
| "learning_rate": 0.00019552783621223436, |
| "loss": 3.1003, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.188, |
| "grad_norm": 2.359375, |
| "learning_rate": 0.0001954240328516277, |
| "loss": 3.2026, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.189, |
| "grad_norm": 1.515625, |
| "learning_rate": 0.0001953190667792947, |
| "loss": 3.1465, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 1.5390625, |
| "learning_rate": 0.00019521293927421388, |
| "loss": 3.1967, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.191, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00019510565162951537, |
| "loss": 3.1578, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.192, |
| "grad_norm": 1.296875, |
| "learning_rate": 0.00019499720515246525, |
| "loss": 3.1388, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.193, |
| "grad_norm": 1.140625, |
| "learning_rate": 0.00019488760116444966, |
| "loss": 3.1873, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.194, |
| "grad_norm": 1.5703125, |
| "learning_rate": 0.0001947768410009586, |
| "loss": 3.1005, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.195, |
| "grad_norm": 1.515625, |
| "learning_rate": 0.00019466492601156966, |
| "loss": 3.0942, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.196, |
| "grad_norm": 1.25, |
| "learning_rate": 0.0001945518575599317, |
| "loss": 3.0119, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.197, |
| "grad_norm": 1.1171875, |
| "learning_rate": 0.00019443763702374812, |
| "loss": 3.0152, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.198, |
| "grad_norm": 1.609375, |
| "learning_rate": 0.0001943222657947601, |
| "loss": 2.9887, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.199, |
| "grad_norm": 1.421875, |
| "learning_rate": 0.00019420574527872968, |
| "loss": 3.0333, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.2, |
| "grad_norm": 1.1171875, |
| "learning_rate": 0.00019408807689542257, |
| "loss": 3.1974, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.201, |
| "grad_norm": 1.4453125, |
| "learning_rate": 0.00019396926207859084, |
| "loss": 3.1598, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.202, |
| "grad_norm": 1.21875, |
| "learning_rate": 0.0001938493022759556, |
| "loss": 3.1167, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.203, |
| "grad_norm": 1.171875, |
| "learning_rate": 0.00019372819894918915, |
| "loss": 3.0386, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.204, |
| "grad_norm": 1.890625, |
| "learning_rate": 0.00019360595357389735, |
| "loss": 3.0719, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.205, |
| "grad_norm": 1.328125, |
| "learning_rate": 0.00019348256763960145, |
| "loss": 3.1225, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.206, |
| "grad_norm": 1.578125, |
| "learning_rate": 0.00019335804264972018, |
| "loss": 3.1224, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.207, |
| "grad_norm": 1.4140625, |
| "learning_rate": 0.00019323238012155123, |
| "loss": 3.0044, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.208, |
| "grad_norm": 1.25, |
| "learning_rate": 0.00019310558158625285, |
| "loss": 3.0157, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.209, |
| "grad_norm": 1.375, |
| "learning_rate": 0.00019297764858882514, |
| "loss": 2.9538, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.21, |
| "grad_norm": 1.375, |
| "learning_rate": 0.00019284858268809137, |
| "loss": 3.0122, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.211, |
| "grad_norm": 1.3359375, |
| "learning_rate": 0.00019271838545667876, |
| "loss": 3.0153, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.212, |
| "grad_norm": 1.015625, |
| "learning_rate": 0.0001925870584809995, |
| "loss": 3.0157, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.213, |
| "grad_norm": 1.484375, |
| "learning_rate": 0.00019245460336123134, |
| "loss": 3.1123, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.214, |
| "grad_norm": 1.1171875, |
| "learning_rate": 0.00019232102171129811, |
| "loss": 3.0119, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.215, |
| "grad_norm": 1.109375, |
| "learning_rate": 0.00019218631515885006, |
| "loss": 3.0785, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.216, |
| "grad_norm": 1.15625, |
| "learning_rate": 0.00019205048534524406, |
| "loss": 2.9962, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.217, |
| "grad_norm": 1.046875, |
| "learning_rate": 0.00019191353392552344, |
| "loss": 2.98, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.218, |
| "grad_norm": 1.2890625, |
| "learning_rate": 0.00019177546256839812, |
| "loss": 3.1094, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.219, |
| "grad_norm": 1.84375, |
| "learning_rate": 0.00019163627295622397, |
| "loss": 3.0632, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.22, |
| "grad_norm": 1.3515625, |
| "learning_rate": 0.0001914959667849825, |
| "loss": 3.0286, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.221, |
| "grad_norm": 1.3359375, |
| "learning_rate": 0.0001913545457642601, |
| "loss": 3.2119, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.222, |
| "grad_norm": 1.203125, |
| "learning_rate": 0.0001912120116172273, |
| "loss": 3.1813, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.223, |
| "grad_norm": 1.1640625, |
| "learning_rate": 0.00019106836608061772, |
| "loss": 3.144, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.224, |
| "grad_norm": 1.2421875, |
| "learning_rate": 0.00019092361090470688, |
| "loss": 3.012, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.225, |
| "grad_norm": 1.1328125, |
| "learning_rate": 0.00019077774785329087, |
| "loss": 3.0367, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.226, |
| "grad_norm": 1.0390625, |
| "learning_rate": 0.000190630778703665, |
| "loss": 3.0175, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.227, |
| "grad_norm": 1.296875, |
| "learning_rate": 0.00019048270524660196, |
| "loss": 2.9683, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.228, |
| "grad_norm": 1.140625, |
| "learning_rate": 0.0001903335292863301, |
| "loss": 3.0164, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.229, |
| "grad_norm": 1.234375, |
| "learning_rate": 0.0001901832526405114, |
| "loss": 2.9548, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.23, |
| "grad_norm": 1.6875, |
| "learning_rate": 0.00019003187714021938, |
| "loss": 2.4954, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.231, |
| "grad_norm": 5.59375, |
| "learning_rate": 0.0001898794046299167, |
| "loss": 2.6911, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.232, |
| "grad_norm": 132.0, |
| "learning_rate": 0.00018972583696743285, |
| "loss": 3.6355, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.233, |
| "grad_norm": 2.875, |
| "learning_rate": 0.0001895711760239413, |
| "loss": 3.0587, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.234, |
| "grad_norm": 1.7734375, |
| "learning_rate": 0.0001894154236839368, |
| "loss": 3.0121, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.235, |
| "grad_norm": 52.0, |
| "learning_rate": 0.00018925858184521256, |
| "loss": 3.0583, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.236, |
| "grad_norm": 2.484375, |
| "learning_rate": 0.0001891006524188368, |
| "loss": 3.1517, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.237, |
| "grad_norm": 1.7578125, |
| "learning_rate": 0.00018894163732912977, |
| "loss": 3.1375, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.238, |
| "grad_norm": 1.59375, |
| "learning_rate": 0.00018878153851364013, |
| "loss": 3.1155, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.239, |
| "grad_norm": 1.546875, |
| "learning_rate": 0.00018862035792312147, |
| "loss": 3.0236, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 1.328125, |
| "learning_rate": 0.0001884580975215084, |
| "loss": 3.2067, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.241, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00018829475928589271, |
| "loss": 3.1324, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.242, |
| "grad_norm": 1.1875, |
| "learning_rate": 0.0001881303452064992, |
| "loss": 3.0433, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.243, |
| "grad_norm": 1.1640625, |
| "learning_rate": 0.00018796485728666165, |
| "loss": 2.9619, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.244, |
| "grad_norm": 1.09375, |
| "learning_rate": 0.00018779829754279805, |
| "loss": 3.052, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.245, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00018763066800438636, |
| "loss": 3.0409, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.246, |
| "grad_norm": 1.6328125, |
| "learning_rate": 0.00018746197071393958, |
| "loss": 2.9298, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.247, |
| "grad_norm": 1.171875, |
| "learning_rate": 0.00018729220772698097, |
| "loss": 3.0326, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.248, |
| "grad_norm": 1.0546875, |
| "learning_rate": 0.00018712138111201895, |
| "loss": 3.0219, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.249, |
| "grad_norm": 1.078125, |
| "learning_rate": 0.0001869494929505219, |
| "loss": 3.0315, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.25, |
| "grad_norm": 11.5, |
| "learning_rate": 0.00018677654533689287, |
| "loss": 2.9556, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.25, |
| "eval_loss": 3.0629920959472656, |
| "eval_runtime": 31.7573, |
| "eval_samples_per_second": 15.744, |
| "eval_steps_per_second": 2.645, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.251, |
| "grad_norm": 1.953125, |
| "learning_rate": 0.00018660254037844388, |
| "loss": 3.0847, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.252, |
| "grad_norm": 1.328125, |
| "learning_rate": 0.0001864274801953705, |
| "loss": 3.0586, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.253, |
| "grad_norm": 1.5390625, |
| "learning_rate": 0.00018625136692072575, |
| "loss": 3.1137, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.254, |
| "grad_norm": 2.4375, |
| "learning_rate": 0.0001860742027003944, |
| "loss": 3.021, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.255, |
| "grad_norm": 1.4375, |
| "learning_rate": 0.00018589598969306645, |
| "loss": 3.0077, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.256, |
| "grad_norm": 1.4453125, |
| "learning_rate": 0.00018571673007021123, |
| "loss": 3.0533, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.257, |
| "grad_norm": 1.453125, |
| "learning_rate": 0.00018553642601605068, |
| "loss": 3.038, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.258, |
| "grad_norm": 2.125, |
| "learning_rate": 0.00018535507972753274, |
| "loss": 2.9965, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.259, |
| "grad_norm": 1.34375, |
| "learning_rate": 0.00018517269341430476, |
| "loss": 3.0632, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.26, |
| "grad_norm": 1.1640625, |
| "learning_rate": 0.00018498926929868642, |
| "loss": 3.1137, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.261, |
| "grad_norm": 23.75, |
| "learning_rate": 0.0001848048096156426, |
| "loss": 2.9885, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.262, |
| "grad_norm": 3.75, |
| "learning_rate": 0.00018461931661275643, |
| "loss": 3.12, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.263, |
| "grad_norm": 2.0625, |
| "learning_rate": 0.00018443279255020152, |
| "loss": 3.1726, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.264, |
| "grad_norm": 1.921875, |
| "learning_rate": 0.00018424523970071477, |
| "loss": 3.2129, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.265, |
| "grad_norm": 1.71875, |
| "learning_rate": 0.00018405666034956844, |
| "loss": 2.9784, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.266, |
| "grad_norm": 71.0, |
| "learning_rate": 0.00018386705679454242, |
| "loss": 3.0557, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.267, |
| "grad_norm": 2.65625, |
| "learning_rate": 0.00018367643134589617, |
| "loss": 3.1084, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.268, |
| "grad_norm": 1.8984375, |
| "learning_rate": 0.00018348478632634066, |
| "loss": 3.0229, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.269, |
| "grad_norm": 3.71875, |
| "learning_rate": 0.00018329212407100994, |
| "loss": 2.9938, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.27, |
| "grad_norm": 2.84375, |
| "learning_rate": 0.00018309844692743283, |
| "loss": 3.029, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.271, |
| "grad_norm": 2.015625, |
| "learning_rate": 0.00018290375725550417, |
| "loss": 2.9886, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.272, |
| "grad_norm": 2.015625, |
| "learning_rate": 0.00018270805742745617, |
| "loss": 3.0823, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.273, |
| "grad_norm": 3.28125, |
| "learning_rate": 0.00018251134982782952, |
| "loss": 3.0021, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.274, |
| "grad_norm": 2.203125, |
| "learning_rate": 0.0001823136368534442, |
| "loss": 2.9724, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.275, |
| "grad_norm": 1.390625, |
| "learning_rate": 0.00018211492091337042, |
| "loss": 2.8975, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.276, |
| "grad_norm": 1.8125, |
| "learning_rate": 0.0001819152044288992, |
| "loss": 2.9522, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.277, |
| "grad_norm": 1.4921875, |
| "learning_rate": 0.00018171448983351284, |
| "loss": 2.9637, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.278, |
| "grad_norm": 1.984375, |
| "learning_rate": 0.00018151277957285543, |
| "loss": 3.0239, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.279, |
| "grad_norm": 1.59375, |
| "learning_rate": 0.00018131007610470276, |
| "loss": 2.9014, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.28, |
| "grad_norm": 1.8828125, |
| "learning_rate": 0.00018110638189893267, |
| "loss": 3.0008, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.281, |
| "grad_norm": 1.4765625, |
| "learning_rate": 0.00018090169943749476, |
| "loss": 3.1132, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.282, |
| "grad_norm": 5.09375, |
| "learning_rate": 0.00018069603121438022, |
| "loss": 3.0337, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.283, |
| "grad_norm": 2.765625, |
| "learning_rate": 0.0001804893797355914, |
| "loss": 3.1402, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.284, |
| "grad_norm": 1.3984375, |
| "learning_rate": 0.00018028174751911146, |
| "loss": 2.9044, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.285, |
| "grad_norm": 6.46875, |
| "learning_rate": 0.00018007313709487334, |
| "loss": 3.0565, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.286, |
| "grad_norm": 2.65625, |
| "learning_rate": 0.00017986355100472928, |
| "loss": 3.1032, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.287, |
| "grad_norm": 3.296875, |
| "learning_rate": 0.00017965299180241963, |
| "loss": 2.9643, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.288, |
| "grad_norm": 2.078125, |
| "learning_rate": 0.00017944146205354182, |
| "loss": 3.0511, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.289, |
| "grad_norm": 1.6328125, |
| "learning_rate": 0.00017922896433551907, |
| "loss": 2.8223, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 1.359375, |
| "learning_rate": 0.00017901550123756906, |
| "loss": 2.9322, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.291, |
| "grad_norm": 13.0, |
| "learning_rate": 0.00017880107536067218, |
| "loss": 3.054, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.292, |
| "grad_norm": 1.7265625, |
| "learning_rate": 0.0001785856893175402, |
| "loss": 3.0824, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.293, |
| "grad_norm": 1.21875, |
| "learning_rate": 0.000178369345732584, |
| "loss": 3.0968, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.294, |
| "grad_norm": 9.25, |
| "learning_rate": 0.00017815204724188187, |
| "loss": 3.138, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.295, |
| "grad_norm": 1.703125, |
| "learning_rate": 0.00017793379649314744, |
| "loss": 3.0392, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.296, |
| "grad_norm": 20.625, |
| "learning_rate": 0.0001777145961456971, |
| "loss": 2.9701, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.297, |
| "grad_norm": 12.0, |
| "learning_rate": 0.00017749444887041799, |
| "loss": 3.0403, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.298, |
| "grad_norm": 2.125, |
| "learning_rate": 0.00017727335734973512, |
| "loss": 2.8949, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.299, |
| "grad_norm": 25.875, |
| "learning_rate": 0.00017705132427757895, |
| "loss": 3.0195, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.3, |
| "grad_norm": 18.25, |
| "learning_rate": 0.00017682835235935236, |
| "loss": 2.9629, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.301, |
| "grad_norm": 3.0, |
| "learning_rate": 0.0001766044443118978, |
| "loss": 3.0801, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.302, |
| "grad_norm": 4.9375, |
| "learning_rate": 0.00017637960286346425, |
| "loss": 2.9368, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.303, |
| "grad_norm": 2.203125, |
| "learning_rate": 0.0001761538307536737, |
| "loss": 2.9964, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.304, |
| "grad_norm": 0.984375, |
| "learning_rate": 0.00017592713073348807, |
| "loss": 2.9558, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.305, |
| "grad_norm": 1.1640625, |
| "learning_rate": 0.00017569950556517566, |
| "loss": 2.9006, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.306, |
| "grad_norm": 1.4140625, |
| "learning_rate": 0.00017547095802227723, |
| "loss": 3.0037, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.307, |
| "grad_norm": 1.0, |
| "learning_rate": 0.00017524149088957245, |
| "loss": 2.8517, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.308, |
| "grad_norm": 0.92578125, |
| "learning_rate": 0.00017501110696304596, |
| "loss": 2.9163, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.309, |
| "grad_norm": 0.98046875, |
| "learning_rate": 0.0001747798090498532, |
| "loss": 2.9563, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.31, |
| "grad_norm": 0.8984375, |
| "learning_rate": 0.00017454759996828623, |
| "loss": 2.8861, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.311, |
| "grad_norm": 0.94140625, |
| "learning_rate": 0.00017431448254773944, |
| "loss": 3.0871, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.312, |
| "grad_norm": 0.98046875, |
| "learning_rate": 0.000174080459628675, |
| "loss": 2.8852, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.313, |
| "grad_norm": 26.0, |
| "learning_rate": 0.00017384553406258842, |
| "loss": 3.0938, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.314, |
| "grad_norm": 1.7265625, |
| "learning_rate": 0.00017360970871197346, |
| "loss": 3.0905, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.315, |
| "grad_norm": 1.140625, |
| "learning_rate": 0.00017337298645028764, |
| "loss": 3.0793, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.316, |
| "grad_norm": 1.25, |
| "learning_rate": 0.00017313537016191706, |
| "loss": 3.0879, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.317, |
| "grad_norm": 1.1875, |
| "learning_rate": 0.00017289686274214118, |
| "loss": 3.1008, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.318, |
| "grad_norm": 1.1015625, |
| "learning_rate": 0.0001726574670970976, |
| "loss": 3.0737, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.319, |
| "grad_norm": 1.0859375, |
| "learning_rate": 0.00017241718614374678, |
| "loss": 2.9513, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 1.046875, |
| "learning_rate": 0.00017217602280983623, |
| "loss": 3.1167, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.321, |
| "grad_norm": 1.0, |
| "learning_rate": 0.0001719339800338651, |
| "loss": 3.0823, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.322, |
| "grad_norm": 1.03125, |
| "learning_rate": 0.0001716910607650483, |
| "loss": 2.9707, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.323, |
| "grad_norm": 0.9921875, |
| "learning_rate": 0.00017144726796328034, |
| "loss": 3.0313, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.324, |
| "grad_norm": 0.93359375, |
| "learning_rate": 0.00017120260459909967, |
| "loss": 2.9786, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.325, |
| "grad_norm": 1.0703125, |
| "learning_rate": 0.0001709570736536521, |
| "loss": 2.9752, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.326, |
| "grad_norm": 0.91796875, |
| "learning_rate": 0.00017071067811865476, |
| "loss": 3.0147, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.327, |
| "grad_norm": 1.1328125, |
| "learning_rate": 0.00017046342099635948, |
| "loss": 3.0414, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.328, |
| "grad_norm": 0.94921875, |
| "learning_rate": 0.00017021530529951625, |
| "loss": 3.0018, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.329, |
| "grad_norm": 1.265625, |
| "learning_rate": 0.00016996633405133655, |
| "loss": 3.0249, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.33, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00016971651028545648, |
| "loss": 2.9958, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.331, |
| "grad_norm": 0.8984375, |
| "learning_rate": 0.00016946583704589973, |
| "loss": 2.9505, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.332, |
| "grad_norm": 1.0703125, |
| "learning_rate": 0.0001692143173870407, |
| "loss": 2.9806, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.333, |
| "grad_norm": 248.0, |
| "learning_rate": 0.000168961954373567, |
| "loss": 2.9853, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.334, |
| "grad_norm": 1.75, |
| "learning_rate": 0.0001687087510804423, |
| "loss": 2.9122, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.335, |
| "grad_norm": 1.0625, |
| "learning_rate": 0.00016845471059286887, |
| "loss": 2.9686, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.336, |
| "grad_norm": 1.4375, |
| "learning_rate": 0.00016819983600624986, |
| "loss": 3.0128, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.337, |
| "grad_norm": 1.4609375, |
| "learning_rate": 0.00016794413042615168, |
| "loss": 2.9692, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.338, |
| "grad_norm": 1.4140625, |
| "learning_rate": 0.00016768759696826608, |
| "loss": 2.9639, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.339, |
| "grad_norm": 1.140625, |
| "learning_rate": 0.00016743023875837233, |
| "loss": 2.8997, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.34, |
| "grad_norm": 1.2421875, |
| "learning_rate": 0.00016717205893229903, |
| "loss": 2.9825, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.341, |
| "grad_norm": 1.1171875, |
| "learning_rate": 0.00016691306063588583, |
| "loss": 2.9442, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.342, |
| "grad_norm": 1.1171875, |
| "learning_rate": 0.00016665324702494524, |
| "loss": 2.9611, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.343, |
| "grad_norm": 1.0, |
| "learning_rate": 0.00016639262126522418, |
| "loss": 3.0123, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.344, |
| "grad_norm": 1.6953125, |
| "learning_rate": 0.00016613118653236518, |
| "loss": 2.9498, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.345, |
| "grad_norm": 1.15625, |
| "learning_rate": 0.00016586894601186805, |
| "loss": 2.8588, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.346, |
| "grad_norm": 0.9609375, |
| "learning_rate": 0.00016560590289905073, |
| "loss": 3.0459, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.347, |
| "grad_norm": 1.0703125, |
| "learning_rate": 0.00016534206039901057, |
| "loss": 2.8838, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.348, |
| "grad_norm": 0.95703125, |
| "learning_rate": 0.0001650774217265851, |
| "loss": 2.9694, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.349, |
| "grad_norm": 0.92578125, |
| "learning_rate": 0.0001648119901063131, |
| "loss": 2.9102, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.35, |
| "grad_norm": 0.921875, |
| "learning_rate": 0.00016454576877239507, |
| "loss": 2.8904, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.351, |
| "grad_norm": 0.85546875, |
| "learning_rate": 0.00016427876096865394, |
| "loss": 2.8798, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.352, |
| "grad_norm": 0.93359375, |
| "learning_rate": 0.00016401096994849557, |
| "loss": 2.9685, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.353, |
| "grad_norm": 0.890625, |
| "learning_rate": 0.000163742398974869, |
| "loss": 2.9498, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.354, |
| "grad_norm": 0.93359375, |
| "learning_rate": 0.00016347305132022677, |
| "loss": 2.9719, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.355, |
| "grad_norm": 0.87890625, |
| "learning_rate": 0.0001632029302664851, |
| "loss": 2.9964, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.356, |
| "grad_norm": 0.921875, |
| "learning_rate": 0.00016293203910498376, |
| "loss": 2.9672, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.357, |
| "grad_norm": 0.87109375, |
| "learning_rate": 0.00016266038113644607, |
| "loss": 2.944, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.358, |
| "grad_norm": 0.87890625, |
| "learning_rate": 0.00016238795967093864, |
| "loss": 2.928, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.359, |
| "grad_norm": 0.91015625, |
| "learning_rate": 0.00016211477802783103, |
| "loss": 2.8984, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.36, |
| "grad_norm": 1.6640625, |
| "learning_rate": 0.0001618408395357554, |
| "loss": 2.807, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.361, |
| "grad_norm": 0.98828125, |
| "learning_rate": 0.0001615661475325658, |
| "loss": 2.9199, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.362, |
| "grad_norm": 1.1171875, |
| "learning_rate": 0.00016129070536529766, |
| "loss": 2.9442, |
| "step": 362 |
| }, |
| { |
| "epoch": 0.363, |
| "grad_norm": 1.1484375, |
| "learning_rate": 0.0001610145163901268, |
| "loss": 2.8548, |
| "step": 363 |
| }, |
| { |
| "epoch": 0.364, |
| "grad_norm": 0.8671875, |
| "learning_rate": 0.00016073758397232868, |
| "loss": 2.9272, |
| "step": 364 |
| }, |
| { |
| "epoch": 0.365, |
| "grad_norm": 0.890625, |
| "learning_rate": 0.0001604599114862375, |
| "loss": 2.9774, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.366, |
| "grad_norm": 0.9609375, |
| "learning_rate": 0.00016018150231520486, |
| "loss": 2.8988, |
| "step": 366 |
| }, |
| { |
| "epoch": 0.367, |
| "grad_norm": 0.8984375, |
| "learning_rate": 0.0001599023598515586, |
| "loss": 2.8573, |
| "step": 367 |
| }, |
| { |
| "epoch": 0.368, |
| "grad_norm": 1.4609375, |
| "learning_rate": 0.0001596224874965616, |
| "loss": 2.86, |
| "step": 368 |
| }, |
| { |
| "epoch": 0.369, |
| "grad_norm": 0.890625, |
| "learning_rate": 0.00015934188866037016, |
| "loss": 2.9432, |
| "step": 369 |
| }, |
| { |
| "epoch": 0.37, |
| "grad_norm": 0.875, |
| "learning_rate": 0.00015906056676199255, |
| "loss": 2.8969, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.371, |
| "grad_norm": 0.91015625, |
| "learning_rate": 0.00015877852522924732, |
| "loss": 2.968, |
| "step": 371 |
| }, |
| { |
| "epoch": 0.372, |
| "grad_norm": 0.86328125, |
| "learning_rate": 0.00015849576749872157, |
| "loss": 2.885, |
| "step": 372 |
| }, |
| { |
| "epoch": 0.373, |
| "grad_norm": 0.82421875, |
| "learning_rate": 0.00015821229701572896, |
| "loss": 2.9352, |
| "step": 373 |
| }, |
| { |
| "epoch": 0.374, |
| "grad_norm": 0.87109375, |
| "learning_rate": 0.0001579281172342679, |
| "loss": 2.7847, |
| "step": 374 |
| }, |
| { |
| "epoch": 0.375, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.00015764323161697935, |
| "loss": 2.9347, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.376, |
| "grad_norm": 0.98046875, |
| "learning_rate": 0.0001573576436351046, |
| "loss": 2.8265, |
| "step": 376 |
| }, |
| { |
| "epoch": 0.377, |
| "grad_norm": 0.921875, |
| "learning_rate": 0.0001570713567684432, |
| "loss": 2.8173, |
| "step": 377 |
| }, |
| { |
| "epoch": 0.378, |
| "grad_norm": 0.9921875, |
| "learning_rate": 0.00015678437450531013, |
| "loss": 2.8425, |
| "step": 378 |
| }, |
| { |
| "epoch": 0.379, |
| "grad_norm": 0.91015625, |
| "learning_rate": 0.0001564967003424938, |
| "loss": 2.7723, |
| "step": 379 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.859375, |
| "learning_rate": 0.00015620833778521307, |
| "loss": 2.9031, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.381, |
| "grad_norm": 0.85546875, |
| "learning_rate": 0.0001559192903470747, |
| "loss": 2.9154, |
| "step": 381 |
| }, |
| { |
| "epoch": 0.382, |
| "grad_norm": 0.86328125, |
| "learning_rate": 0.0001556295615500305, |
| "loss": 2.8226, |
| "step": 382 |
| }, |
| { |
| "epoch": 0.383, |
| "grad_norm": 1.015625, |
| "learning_rate": 0.00015533915492433443, |
| "loss": 2.8911, |
| "step": 383 |
| }, |
| { |
| "epoch": 0.384, |
| "grad_norm": 0.8828125, |
| "learning_rate": 0.00015504807400849958, |
| "loss": 2.8114, |
| "step": 384 |
| }, |
| { |
| "epoch": 0.385, |
| "grad_norm": 0.890625, |
| "learning_rate": 0.00015475632234925504, |
| "loss": 2.8762, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.386, |
| "grad_norm": 0.828125, |
| "learning_rate": 0.00015446390350150273, |
| "loss": 2.8918, |
| "step": 386 |
| }, |
| { |
| "epoch": 0.387, |
| "grad_norm": 0.9296875, |
| "learning_rate": 0.000154170821028274, |
| "loss": 2.9433, |
| "step": 387 |
| }, |
| { |
| "epoch": 0.388, |
| "grad_norm": 0.84375, |
| "learning_rate": 0.0001538770785006863, |
| "loss": 2.871, |
| "step": 388 |
| }, |
| { |
| "epoch": 0.389, |
| "grad_norm": 0.90625, |
| "learning_rate": 0.00015358267949789966, |
| "loss": 2.8932, |
| "step": 389 |
| }, |
| { |
| "epoch": 0.39, |
| "grad_norm": 0.86328125, |
| "learning_rate": 0.000153287627607073, |
| "loss": 2.8251, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.391, |
| "grad_norm": 1.21875, |
| "learning_rate": 0.0001529919264233205, |
| "loss": 2.9062, |
| "step": 391 |
| }, |
| { |
| "epoch": 0.392, |
| "grad_norm": 1.0625, |
| "learning_rate": 0.00015269557954966778, |
| "loss": 2.8995, |
| "step": 392 |
| }, |
| { |
| "epoch": 0.393, |
| "grad_norm": 0.796875, |
| "learning_rate": 0.00015239859059700794, |
| "loss": 2.7726, |
| "step": 393 |
| }, |
| { |
| "epoch": 0.394, |
| "grad_norm": 1.0625, |
| "learning_rate": 0.00015210096318405767, |
| "loss": 2.7649, |
| "step": 394 |
| }, |
| { |
| "epoch": 0.395, |
| "grad_norm": 0.92578125, |
| "learning_rate": 0.00015180270093731303, |
| "loss": 2.778, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.396, |
| "grad_norm": 0.94921875, |
| "learning_rate": 0.00015150380749100545, |
| "loss": 2.8322, |
| "step": 396 |
| }, |
| { |
| "epoch": 0.397, |
| "grad_norm": 0.91796875, |
| "learning_rate": 0.00015120428648705717, |
| "loss": 2.8831, |
| "step": 397 |
| }, |
| { |
| "epoch": 0.398, |
| "grad_norm": 0.8671875, |
| "learning_rate": 0.00015090414157503714, |
| "loss": 2.8862, |
| "step": 398 |
| }, |
| { |
| "epoch": 0.399, |
| "grad_norm": 0.84765625, |
| "learning_rate": 0.00015060337641211637, |
| "loss": 2.8219, |
| "step": 399 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 0.7890625, |
| "learning_rate": 0.00015030199466302353, |
| "loss": 2.8894, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.401, |
| "grad_norm": 0.83984375, |
| "learning_rate": 0.00015000000000000001, |
| "loss": 2.8594, |
| "step": 401 |
| }, |
| { |
| "epoch": 0.402, |
| "grad_norm": 0.875, |
| "learning_rate": 0.00014969739610275556, |
| "loss": 2.7943, |
| "step": 402 |
| }, |
| { |
| "epoch": 0.403, |
| "grad_norm": 0.890625, |
| "learning_rate": 0.0001493941866584231, |
| "loss": 2.9105, |
| "step": 403 |
| }, |
| { |
| "epoch": 0.404, |
| "grad_norm": 1.3046875, |
| "learning_rate": 0.00014909037536151409, |
| "loss": 2.9341, |
| "step": 404 |
| }, |
| { |
| "epoch": 0.405, |
| "grad_norm": 0.984375, |
| "learning_rate": 0.0001487859659138733, |
| "loss": 2.8726, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.406, |
| "grad_norm": 0.796875, |
| "learning_rate": 0.00014848096202463372, |
| "loss": 2.9266, |
| "step": 406 |
| }, |
| { |
| "epoch": 0.407, |
| "grad_norm": 12.5, |
| "learning_rate": 0.00014817536741017152, |
| "loss": 2.8098, |
| "step": 407 |
| }, |
| { |
| "epoch": 0.408, |
| "grad_norm": 1.21875, |
| "learning_rate": 0.0001478691857940607, |
| "loss": 2.8836, |
| "step": 408 |
| }, |
| { |
| "epoch": 0.409, |
| "grad_norm": 1.1328125, |
| "learning_rate": 0.00014756242090702756, |
| "loss": 2.8176, |
| "step": 409 |
| }, |
| { |
| "epoch": 0.41, |
| "grad_norm": 1.265625, |
| "learning_rate": 0.00014725507648690543, |
| "loss": 2.8858, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.411, |
| "grad_norm": 1.1328125, |
| "learning_rate": 0.00014694715627858908, |
| "loss": 2.8898, |
| "step": 411 |
| }, |
| { |
| "epoch": 0.412, |
| "grad_norm": 1.078125, |
| "learning_rate": 0.00014663866403398913, |
| "loss": 2.8529, |
| "step": 412 |
| }, |
| { |
| "epoch": 0.413, |
| "grad_norm": 1.0625, |
| "learning_rate": 0.00014632960351198618, |
| "loss": 2.8151, |
| "step": 413 |
| }, |
| { |
| "epoch": 0.414, |
| "grad_norm": 1.0078125, |
| "learning_rate": 0.00014601997847838518, |
| "loss": 2.9201, |
| "step": 414 |
| }, |
| { |
| "epoch": 0.415, |
| "grad_norm": 1.0078125, |
| "learning_rate": 0.00014570979270586945, |
| "loss": 2.7901, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.416, |
| "grad_norm": 1.1015625, |
| "learning_rate": 0.00014539904997395468, |
| "loss": 2.8675, |
| "step": 416 |
| }, |
| { |
| "epoch": 0.417, |
| "grad_norm": 0.9453125, |
| "learning_rate": 0.00014508775406894307, |
| "loss": 2.7743, |
| "step": 417 |
| }, |
| { |
| "epoch": 0.418, |
| "grad_norm": 0.98046875, |
| "learning_rate": 0.00014477590878387696, |
| "loss": 2.8796, |
| "step": 418 |
| }, |
| { |
| "epoch": 0.419, |
| "grad_norm": 0.94921875, |
| "learning_rate": 0.00014446351791849276, |
| "loss": 2.9855, |
| "step": 419 |
| }, |
| { |
| "epoch": 0.42, |
| "grad_norm": 0.95703125, |
| "learning_rate": 0.00014415058527917452, |
| "loss": 2.8645, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.421, |
| "grad_norm": 0.91015625, |
| "learning_rate": 0.00014383711467890774, |
| "loss": 2.7821, |
| "step": 421 |
| }, |
| { |
| "epoch": 0.422, |
| "grad_norm": 5.53125, |
| "learning_rate": 0.00014352310993723277, |
| "loss": 2.9062, |
| "step": 422 |
| }, |
| { |
| "epoch": 0.423, |
| "grad_norm": 0.96484375, |
| "learning_rate": 0.00014320857488019824, |
| "loss": 2.8564, |
| "step": 423 |
| }, |
| { |
| "epoch": 0.424, |
| "grad_norm": 0.84765625, |
| "learning_rate": 0.0001428935133403146, |
| "loss": 2.855, |
| "step": 424 |
| }, |
| { |
| "epoch": 0.425, |
| "grad_norm": 0.91796875, |
| "learning_rate": 0.00014257792915650728, |
| "loss": 2.907, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.426, |
| "grad_norm": 1.0546875, |
| "learning_rate": 0.00014226182617406996, |
| "loss": 2.8238, |
| "step": 426 |
| }, |
| { |
| "epoch": 0.427, |
| "grad_norm": 1.09375, |
| "learning_rate": 0.00014194520824461771, |
| "loss": 2.8523, |
| "step": 427 |
| }, |
| { |
| "epoch": 0.428, |
| "grad_norm": 1.1015625, |
| "learning_rate": 0.00014162807922604012, |
| "loss": 2.91, |
| "step": 428 |
| }, |
| { |
| "epoch": 0.429, |
| "grad_norm": 1.2109375, |
| "learning_rate": 0.0001413104429824542, |
| "loss": 2.9161, |
| "step": 429 |
| }, |
| { |
| "epoch": 0.43, |
| "grad_norm": 0.8515625, |
| "learning_rate": 0.00014099230338415728, |
| "loss": 2.9304, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.431, |
| "grad_norm": 0.75, |
| "learning_rate": 0.00014067366430758004, |
| "loss": 2.8266, |
| "step": 431 |
| }, |
| { |
| "epoch": 0.432, |
| "grad_norm": 0.8203125, |
| "learning_rate": 0.00014035452963523902, |
| "loss": 2.8303, |
| "step": 432 |
| }, |
| { |
| "epoch": 0.433, |
| "grad_norm": 0.7890625, |
| "learning_rate": 0.00014003490325568954, |
| "loss": 2.8311, |
| "step": 433 |
| }, |
| { |
| "epoch": 0.434, |
| "grad_norm": 0.7265625, |
| "learning_rate": 0.00013971478906347806, |
| "loss": 2.826, |
| "step": 434 |
| }, |
| { |
| "epoch": 0.435, |
| "grad_norm": 0.8828125, |
| "learning_rate": 0.00013939419095909512, |
| "loss": 2.6267, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.436, |
| "grad_norm": 0.77734375, |
| "learning_rate": 0.00013907311284892736, |
| "loss": 2.7203, |
| "step": 436 |
| }, |
| { |
| "epoch": 0.437, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.0001387515586452103, |
| "loss": 2.8801, |
| "step": 437 |
| }, |
| { |
| "epoch": 0.438, |
| "grad_norm": 2.421875, |
| "learning_rate": 0.00013842953226598037, |
| "loss": 2.9568, |
| "step": 438 |
| }, |
| { |
| "epoch": 0.439, |
| "grad_norm": 0.9375, |
| "learning_rate": 0.00013810703763502744, |
| "loss": 2.758, |
| "step": 439 |
| }, |
| { |
| "epoch": 0.44, |
| "grad_norm": 1.3046875, |
| "learning_rate": 0.00013778407868184672, |
| "loss": 2.8374, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.441, |
| "grad_norm": 0.91796875, |
| "learning_rate": 0.00013746065934159123, |
| "loss": 2.7894, |
| "step": 441 |
| }, |
| { |
| "epoch": 0.442, |
| "grad_norm": 0.87109375, |
| "learning_rate": 0.00013713678355502351, |
| "loss": 2.855, |
| "step": 442 |
| }, |
| { |
| "epoch": 0.443, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.00013681245526846783, |
| "loss": 2.777, |
| "step": 443 |
| }, |
| { |
| "epoch": 0.444, |
| "grad_norm": 0.875, |
| "learning_rate": 0.00013648767843376196, |
| "loss": 2.8408, |
| "step": 444 |
| }, |
| { |
| "epoch": 0.445, |
| "grad_norm": 0.80078125, |
| "learning_rate": 0.00013616245700820922, |
| "loss": 2.8602, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.446, |
| "grad_norm": 0.8359375, |
| "learning_rate": 0.00013583679495453, |
| "loss": 2.9452, |
| "step": 446 |
| }, |
| { |
| "epoch": 0.447, |
| "grad_norm": 0.83203125, |
| "learning_rate": 0.0001355106962408137, |
| "loss": 2.799, |
| "step": 447 |
| }, |
| { |
| "epoch": 0.448, |
| "grad_norm": 0.78125, |
| "learning_rate": 0.00013518416484047018, |
| "loss": 2.8164, |
| "step": 448 |
| }, |
| { |
| "epoch": 0.449, |
| "grad_norm": 0.78515625, |
| "learning_rate": 0.00013485720473218154, |
| "loss": 2.9159, |
| "step": 449 |
| }, |
| { |
| "epoch": 0.45, |
| "grad_norm": 0.77734375, |
| "learning_rate": 0.00013452981989985348, |
| "loss": 2.8874, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.451, |
| "grad_norm": 0.984375, |
| "learning_rate": 0.00013420201433256689, |
| "loss": 2.8846, |
| "step": 451 |
| }, |
| { |
| "epoch": 0.452, |
| "grad_norm": 0.78515625, |
| "learning_rate": 0.00013387379202452917, |
| "loss": 2.8474, |
| "step": 452 |
| }, |
| { |
| "epoch": 0.453, |
| "grad_norm": 0.86328125, |
| "learning_rate": 0.00013354515697502553, |
| "loss": 2.8697, |
| "step": 453 |
| }, |
| { |
| "epoch": 0.454, |
| "grad_norm": 2.875, |
| "learning_rate": 0.00013321611318837032, |
| "loss": 2.8103, |
| "step": 454 |
| }, |
| { |
| "epoch": 0.455, |
| "grad_norm": 0.88671875, |
| "learning_rate": 0.00013288666467385833, |
| "loss": 2.7893, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.456, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.00013255681544571568, |
| "loss": 2.8302, |
| "step": 456 |
| }, |
| { |
| "epoch": 0.457, |
| "grad_norm": 0.76953125, |
| "learning_rate": 0.00013222656952305113, |
| "loss": 2.8056, |
| "step": 457 |
| }, |
| { |
| "epoch": 0.458, |
| "grad_norm": 0.7734375, |
| "learning_rate": 0.00013189593092980702, |
| "loss": 2.8463, |
| "step": 458 |
| }, |
| { |
| "epoch": 0.459, |
| "grad_norm": 0.75, |
| "learning_rate": 0.00013156490369471027, |
| "loss": 2.7264, |
| "step": 459 |
| }, |
| { |
| "epoch": 0.46, |
| "grad_norm": 0.75390625, |
| "learning_rate": 0.00013123349185122327, |
| "loss": 2.7938, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.461, |
| "grad_norm": 0.76953125, |
| "learning_rate": 0.00013090169943749476, |
| "loss": 2.7834, |
| "step": 461 |
| }, |
| { |
| "epoch": 0.462, |
| "grad_norm": 0.7265625, |
| "learning_rate": 0.00013056953049631057, |
| "loss": 2.8079, |
| "step": 462 |
| }, |
| { |
| "epoch": 0.463, |
| "grad_norm": 0.73046875, |
| "learning_rate": 0.00013023698907504446, |
| "loss": 2.7598, |
| "step": 463 |
| }, |
| { |
| "epoch": 0.464, |
| "grad_norm": 0.71875, |
| "learning_rate": 0.00012990407922560868, |
| "loss": 2.8624, |
| "step": 464 |
| }, |
| { |
| "epoch": 0.465, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.00012957080500440468, |
| "loss": 2.8337, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.466, |
| "grad_norm": 0.71484375, |
| "learning_rate": 0.00012923717047227368, |
| "loss": 2.8641, |
| "step": 466 |
| }, |
| { |
| "epoch": 0.467, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.00012890317969444716, |
| "loss": 2.832, |
| "step": 467 |
| }, |
| { |
| "epoch": 0.468, |
| "grad_norm": 0.765625, |
| "learning_rate": 0.00012856883674049736, |
| "loss": 2.9126, |
| "step": 468 |
| }, |
| { |
| "epoch": 0.469, |
| "grad_norm": 3.4375, |
| "learning_rate": 0.00012823414568428768, |
| "loss": 2.9386, |
| "step": 469 |
| }, |
| { |
| "epoch": 0.47, |
| "grad_norm": 0.81640625, |
| "learning_rate": 0.00012789911060392294, |
| "loss": 2.7886, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.471, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.0001275637355816999, |
| "loss": 2.9174, |
| "step": 471 |
| }, |
| { |
| "epoch": 0.472, |
| "grad_norm": 0.7890625, |
| "learning_rate": 0.00012722802470405744, |
| "loss": 2.7964, |
| "step": 472 |
| }, |
| { |
| "epoch": 0.473, |
| "grad_norm": 0.83984375, |
| "learning_rate": 0.00012689198206152657, |
| "loss": 2.779, |
| "step": 473 |
| }, |
| { |
| "epoch": 0.474, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.00012655561174868088, |
| "loss": 2.8909, |
| "step": 474 |
| }, |
| { |
| "epoch": 0.475, |
| "grad_norm": 0.7890625, |
| "learning_rate": 0.00012621891786408648, |
| "loss": 2.8705, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.476, |
| "grad_norm": 0.83203125, |
| "learning_rate": 0.00012588190451025207, |
| "loss": 2.9862, |
| "step": 476 |
| }, |
| { |
| "epoch": 0.477, |
| "grad_norm": 0.74609375, |
| "learning_rate": 0.00012554457579357905, |
| "loss": 2.8599, |
| "step": 477 |
| }, |
| { |
| "epoch": 0.478, |
| "grad_norm": 0.73828125, |
| "learning_rate": 0.0001252069358243114, |
| "loss": 2.9299, |
| "step": 478 |
| }, |
| { |
| "epoch": 0.479, |
| "grad_norm": 0.78125, |
| "learning_rate": 0.0001248689887164855, |
| "loss": 2.8572, |
| "step": 479 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.78515625, |
| "learning_rate": 0.00012453073858788026, |
| "loss": 2.8889, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.481, |
| "grad_norm": 0.7109375, |
| "learning_rate": 0.00012419218955996676, |
| "loss": 2.8363, |
| "step": 481 |
| }, |
| { |
| "epoch": 0.482, |
| "grad_norm": 0.7578125, |
| "learning_rate": 0.0001238533457578581, |
| "loss": 2.8624, |
| "step": 482 |
| }, |
| { |
| "epoch": 0.483, |
| "grad_norm": 0.7734375, |
| "learning_rate": 0.000123514211310259, |
| "loss": 2.8517, |
| "step": 483 |
| }, |
| { |
| "epoch": 0.484, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.00012317479034941573, |
| "loss": 2.8247, |
| "step": 484 |
| }, |
| { |
| "epoch": 0.485, |
| "grad_norm": 2.140625, |
| "learning_rate": 0.00012283508701106557, |
| "loss": 2.8114, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.486, |
| "grad_norm": 0.8515625, |
| "learning_rate": 0.0001224951054343865, |
| "loss": 2.7308, |
| "step": 486 |
| }, |
| { |
| "epoch": 0.487, |
| "grad_norm": 0.76171875, |
| "learning_rate": 0.00012215484976194676, |
| "loss": 2.8445, |
| "step": 487 |
| }, |
| { |
| "epoch": 0.488, |
| "grad_norm": 0.8828125, |
| "learning_rate": 0.00012181432413965428, |
| "loss": 2.933, |
| "step": 488 |
| }, |
| { |
| "epoch": 0.489, |
| "grad_norm": 0.72265625, |
| "learning_rate": 0.00012147353271670634, |
| "loss": 2.8725, |
| "step": 489 |
| }, |
| { |
| "epoch": 0.49, |
| "grad_norm": 0.796875, |
| "learning_rate": 0.00012113247964553888, |
| "loss": 2.7981, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.491, |
| "grad_norm": 0.74609375, |
| "learning_rate": 0.00012079116908177593, |
| "loss": 2.8803, |
| "step": 491 |
| }, |
| { |
| "epoch": 0.492, |
| "grad_norm": 0.73046875, |
| "learning_rate": 0.00012044960518417903, |
| "loss": 2.7637, |
| "step": 492 |
| }, |
| { |
| "epoch": 0.493, |
| "grad_norm": 0.75, |
| "learning_rate": 0.00012010779211459648, |
| "loss": 2.888, |
| "step": 493 |
| }, |
| { |
| "epoch": 0.494, |
| "grad_norm": 0.7265625, |
| "learning_rate": 0.00011976573403791262, |
| "loss": 2.815, |
| "step": 494 |
| }, |
| { |
| "epoch": 0.495, |
| "grad_norm": 0.77734375, |
| "learning_rate": 0.0001194234351219972, |
| "loss": 2.7684, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.496, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.00011908089953765449, |
| "loss": 2.7292, |
| "step": 496 |
| }, |
| { |
| "epoch": 0.497, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.00011873813145857249, |
| "loss": 2.7949, |
| "step": 497 |
| }, |
| { |
| "epoch": 0.498, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.00011839513506127203, |
| "loss": 2.7681, |
| "step": 498 |
| }, |
| { |
| "epoch": 0.499, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.00011805191452505602, |
| "loss": 2.6686, |
| "step": 499 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 1.2578125, |
| "learning_rate": 0.00011770847403195834, |
| "loss": 2.7635, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.5, |
| "eval_loss": 2.7894091606140137, |
| "eval_runtime": 31.8024, |
| "eval_samples_per_second": 15.722, |
| "eval_steps_per_second": 2.641, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.501, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.00011736481776669306, |
| "loss": 2.7591, |
| "step": 501 |
| }, |
| { |
| "epoch": 0.502, |
| "grad_norm": 0.73828125, |
| "learning_rate": 0.00011702094991660326, |
| "loss": 2.8107, |
| "step": 502 |
| }, |
| { |
| "epoch": 0.503, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.00011667687467161024, |
| "loss": 2.8147, |
| "step": 503 |
| }, |
| { |
| "epoch": 0.504, |
| "grad_norm": 0.76171875, |
| "learning_rate": 0.00011633259622416224, |
| "loss": 2.8342, |
| "step": 504 |
| }, |
| { |
| "epoch": 0.505, |
| "grad_norm": 0.79296875, |
| "learning_rate": 0.0001159881187691835, |
| "loss": 2.8037, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.506, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.0001156434465040231, |
| "loss": 2.8258, |
| "step": 506 |
| }, |
| { |
| "epoch": 0.507, |
| "grad_norm": 0.84375, |
| "learning_rate": 0.00011529858362840382, |
| "loss": 2.8476, |
| "step": 507 |
| }, |
| { |
| "epoch": 0.508, |
| "grad_norm": 0.74609375, |
| "learning_rate": 0.00011495353434437098, |
| "loss": 2.7726, |
| "step": 508 |
| }, |
| { |
| "epoch": 0.509, |
| "grad_norm": 0.7890625, |
| "learning_rate": 0.00011460830285624118, |
| "loss": 2.7936, |
| "step": 509 |
| }, |
| { |
| "epoch": 0.51, |
| "grad_norm": 0.71875, |
| "learning_rate": 0.00011426289337055119, |
| "loss": 2.7585, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.511, |
| "grad_norm": 0.703125, |
| "learning_rate": 0.00011391731009600654, |
| "loss": 2.7722, |
| "step": 511 |
| }, |
| { |
| "epoch": 0.512, |
| "grad_norm": 0.81640625, |
| "learning_rate": 0.00011357155724343045, |
| "loss": 2.8518, |
| "step": 512 |
| }, |
| { |
| "epoch": 0.513, |
| "grad_norm": 0.6484375, |
| "learning_rate": 0.00011322563902571226, |
| "loss": 2.8035, |
| "step": 513 |
| }, |
| { |
| "epoch": 0.514, |
| "grad_norm": 0.72265625, |
| "learning_rate": 0.0001128795596577563, |
| "loss": 2.8544, |
| "step": 514 |
| }, |
| { |
| "epoch": 0.515, |
| "grad_norm": 0.73828125, |
| "learning_rate": 0.00011253332335643043, |
| "loss": 2.8212, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.516, |
| "grad_norm": 2.21875, |
| "learning_rate": 0.00011218693434051475, |
| "loss": 2.7369, |
| "step": 516 |
| }, |
| { |
| "epoch": 0.517, |
| "grad_norm": 0.8203125, |
| "learning_rate": 0.00011184039683065013, |
| "loss": 2.7366, |
| "step": 517 |
| }, |
| { |
| "epoch": 0.518, |
| "grad_norm": 0.73046875, |
| "learning_rate": 0.00011149371504928668, |
| "loss": 2.8155, |
| "step": 518 |
| }, |
| { |
| "epoch": 0.519, |
| "grad_norm": 0.77734375, |
| "learning_rate": 0.00011114689322063255, |
| "loss": 2.7464, |
| "step": 519 |
| }, |
| { |
| "epoch": 0.52, |
| "grad_norm": 0.8203125, |
| "learning_rate": 0.0001107999355706023, |
| "loss": 2.796, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.521, |
| "grad_norm": 0.71875, |
| "learning_rate": 0.00011045284632676536, |
| "loss": 2.7958, |
| "step": 521 |
| }, |
| { |
| "epoch": 0.522, |
| "grad_norm": 0.703125, |
| "learning_rate": 0.00011010562971829463, |
| "loss": 2.7453, |
| "step": 522 |
| }, |
| { |
| "epoch": 0.523, |
| "grad_norm": 0.7421875, |
| "learning_rate": 0.00010975828997591495, |
| "loss": 2.7748, |
| "step": 523 |
| }, |
| { |
| "epoch": 0.524, |
| "grad_norm": 0.65625, |
| "learning_rate": 0.00010941083133185146, |
| "loss": 2.8185, |
| "step": 524 |
| }, |
| { |
| "epoch": 0.525, |
| "grad_norm": 0.7890625, |
| "learning_rate": 0.00010906325801977804, |
| "loss": 2.7946, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.526, |
| "grad_norm": 0.71875, |
| "learning_rate": 0.00010871557427476583, |
| "loss": 2.8366, |
| "step": 526 |
| }, |
| { |
| "epoch": 0.527, |
| "grad_norm": 0.63671875, |
| "learning_rate": 0.00010836778433323158, |
| "loss": 2.81, |
| "step": 527 |
| }, |
| { |
| "epoch": 0.528, |
| "grad_norm": 0.7109375, |
| "learning_rate": 0.00010801989243288589, |
| "loss": 2.7869, |
| "step": 528 |
| }, |
| { |
| "epoch": 0.529, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.00010767190281268187, |
| "loss": 2.7461, |
| "step": 529 |
| }, |
| { |
| "epoch": 0.53, |
| "grad_norm": 0.6875, |
| "learning_rate": 0.00010732381971276318, |
| "loss": 2.7465, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.531, |
| "grad_norm": 0.91796875, |
| "learning_rate": 0.00010697564737441252, |
| "loss": 2.7434, |
| "step": 531 |
| }, |
| { |
| "epoch": 0.532, |
| "grad_norm": 6.3125, |
| "learning_rate": 0.00010662739004000005, |
| "loss": 2.7999, |
| "step": 532 |
| }, |
| { |
| "epoch": 0.533, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.00010627905195293135, |
| "loss": 2.7405, |
| "step": 533 |
| }, |
| { |
| "epoch": 0.534, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.00010593063735759618, |
| "loss": 2.8709, |
| "step": 534 |
| }, |
| { |
| "epoch": 0.535, |
| "grad_norm": 0.66796875, |
| "learning_rate": 0.00010558215049931638, |
| "loss": 2.9645, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.536, |
| "grad_norm": 0.7265625, |
| "learning_rate": 0.0001052335956242944, |
| "loss": 2.7778, |
| "step": 536 |
| }, |
| { |
| "epoch": 0.537, |
| "grad_norm": 0.66796875, |
| "learning_rate": 0.00010488497697956135, |
| "loss": 2.7532, |
| "step": 537 |
| }, |
| { |
| "epoch": 0.538, |
| "grad_norm": 0.8203125, |
| "learning_rate": 0.00010453629881292538, |
| "loss": 2.8057, |
| "step": 538 |
| }, |
| { |
| "epoch": 0.539, |
| "grad_norm": 0.70703125, |
| "learning_rate": 0.00010418756537291996, |
| "loss": 2.75, |
| "step": 539 |
| }, |
| { |
| "epoch": 0.54, |
| "grad_norm": 0.7109375, |
| "learning_rate": 0.00010383878090875201, |
| "loss": 2.8593, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.541, |
| "grad_norm": 0.6640625, |
| "learning_rate": 0.00010348994967025012, |
| "loss": 2.7264, |
| "step": 541 |
| }, |
| { |
| "epoch": 0.542, |
| "grad_norm": 0.703125, |
| "learning_rate": 0.00010314107590781284, |
| "loss": 2.8229, |
| "step": 542 |
| }, |
| { |
| "epoch": 0.543, |
| "grad_norm": 0.68359375, |
| "learning_rate": 0.0001027921638723569, |
| "loss": 2.7506, |
| "step": 543 |
| }, |
| { |
| "epoch": 0.544, |
| "grad_norm": 0.65625, |
| "learning_rate": 0.00010244321781526533, |
| "loss": 2.7689, |
| "step": 544 |
| }, |
| { |
| "epoch": 0.545, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.0001020942419883357, |
| "loss": 2.766, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.546, |
| "grad_norm": 0.69140625, |
| "learning_rate": 0.00010174524064372837, |
| "loss": 2.8915, |
| "step": 546 |
| }, |
| { |
| "epoch": 0.547, |
| "grad_norm": 2.625, |
| "learning_rate": 0.00010139621803391455, |
| "loss": 2.7575, |
| "step": 547 |
| }, |
| { |
| "epoch": 0.548, |
| "grad_norm": 0.7578125, |
| "learning_rate": 0.00010104717841162458, |
| "loss": 2.7613, |
| "step": 548 |
| }, |
| { |
| "epoch": 0.549, |
| "grad_norm": 0.6875, |
| "learning_rate": 0.00010069812602979615, |
| "loss": 2.8354, |
| "step": 549 |
| }, |
| { |
| "epoch": 0.55, |
| "grad_norm": 0.640625, |
| "learning_rate": 0.00010034906514152238, |
| "loss": 2.815, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.551, |
| "grad_norm": 0.78515625, |
| "learning_rate": 0.0001, |
| "loss": 2.9268, |
| "step": 551 |
| }, |
| { |
| "epoch": 0.552, |
| "grad_norm": 0.7109375, |
| "learning_rate": 9.965093485847767e-05, |
| "loss": 2.7682, |
| "step": 552 |
| }, |
| { |
| "epoch": 0.553, |
| "grad_norm": 0.6484375, |
| "learning_rate": 9.930187397020386e-05, |
| "loss": 2.8294, |
| "step": 553 |
| }, |
| { |
| "epoch": 0.554, |
| "grad_norm": 0.734375, |
| "learning_rate": 9.895282158837545e-05, |
| "loss": 2.7828, |
| "step": 554 |
| }, |
| { |
| "epoch": 0.555, |
| "grad_norm": 0.6953125, |
| "learning_rate": 9.860378196608549e-05, |
| "loss": 2.7539, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.556, |
| "grad_norm": 0.6953125, |
| "learning_rate": 9.825475935627165e-05, |
| "loss": 2.7614, |
| "step": 556 |
| }, |
| { |
| "epoch": 0.557, |
| "grad_norm": 0.73828125, |
| "learning_rate": 9.790575801166432e-05, |
| "loss": 2.7728, |
| "step": 557 |
| }, |
| { |
| "epoch": 0.558, |
| "grad_norm": 0.7265625, |
| "learning_rate": 9.755678218473469e-05, |
| "loss": 2.7955, |
| "step": 558 |
| }, |
| { |
| "epoch": 0.559, |
| "grad_norm": 0.6796875, |
| "learning_rate": 9.720783612764314e-05, |
| "loss": 2.8389, |
| "step": 559 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 0.70703125, |
| "learning_rate": 9.685892409218717e-05, |
| "loss": 2.7221, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.561, |
| "grad_norm": 0.69140625, |
| "learning_rate": 9.651005032974994e-05, |
| "loss": 2.7392, |
| "step": 561 |
| }, |
| { |
| "epoch": 0.562, |
| "grad_norm": 0.65625, |
| "learning_rate": 9.616121909124801e-05, |
| "loss": 2.7522, |
| "step": 562 |
| }, |
| { |
| "epoch": 0.563, |
| "grad_norm": 4.125, |
| "learning_rate": 9.581243462708006e-05, |
| "loss": 2.7441, |
| "step": 563 |
| }, |
| { |
| "epoch": 0.564, |
| "grad_norm": 0.81640625, |
| "learning_rate": 9.546370118707463e-05, |
| "loss": 2.7218, |
| "step": 564 |
| }, |
| { |
| "epoch": 0.565, |
| "grad_norm": 0.73046875, |
| "learning_rate": 9.511502302043868e-05, |
| "loss": 2.7805, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.566, |
| "grad_norm": 0.67578125, |
| "learning_rate": 9.476640437570562e-05, |
| "loss": 2.7791, |
| "step": 566 |
| }, |
| { |
| "epoch": 0.567, |
| "grad_norm": 0.69921875, |
| "learning_rate": 9.441784950068362e-05, |
| "loss": 2.8229, |
| "step": 567 |
| }, |
| { |
| "epoch": 0.568, |
| "grad_norm": 0.71875, |
| "learning_rate": 9.406936264240386e-05, |
| "loss": 2.8902, |
| "step": 568 |
| }, |
| { |
| "epoch": 0.569, |
| "grad_norm": 0.71484375, |
| "learning_rate": 9.372094804706867e-05, |
| "loss": 2.8009, |
| "step": 569 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 0.62890625, |
| "learning_rate": 9.337260996000002e-05, |
| "loss": 2.7867, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.571, |
| "grad_norm": 0.71875, |
| "learning_rate": 9.302435262558747e-05, |
| "loss": 2.7513, |
| "step": 571 |
| }, |
| { |
| "epoch": 0.572, |
| "grad_norm": 0.6875, |
| "learning_rate": 9.267618028723686e-05, |
| "loss": 2.8066, |
| "step": 572 |
| }, |
| { |
| "epoch": 0.573, |
| "grad_norm": 0.6484375, |
| "learning_rate": 9.232809718731814e-05, |
| "loss": 2.818, |
| "step": 573 |
| }, |
| { |
| "epoch": 0.574, |
| "grad_norm": 0.69921875, |
| "learning_rate": 9.198010756711412e-05, |
| "loss": 2.6698, |
| "step": 574 |
| }, |
| { |
| "epoch": 0.575, |
| "grad_norm": 0.7109375, |
| "learning_rate": 9.163221566676847e-05, |
| "loss": 2.7901, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.576, |
| "grad_norm": 2.0625, |
| "learning_rate": 9.128442572523417e-05, |
| "loss": 2.8648, |
| "step": 576 |
| }, |
| { |
| "epoch": 0.577, |
| "grad_norm": 0.78515625, |
| "learning_rate": 9.093674198022201e-05, |
| "loss": 2.8698, |
| "step": 577 |
| }, |
| { |
| "epoch": 0.578, |
| "grad_norm": 0.66796875, |
| "learning_rate": 9.058916866814858e-05, |
| "loss": 2.8594, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.579, |
| "grad_norm": 3.65625, |
| "learning_rate": 9.024171002408506e-05, |
| "loss": 2.788, |
| "step": 579 |
| }, |
| { |
| "epoch": 0.58, |
| "grad_norm": 0.8125, |
| "learning_rate": 8.989437028170537e-05, |
| "loss": 2.7126, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.581, |
| "grad_norm": 0.734375, |
| "learning_rate": 8.954715367323468e-05, |
| "loss": 2.7795, |
| "step": 581 |
| }, |
| { |
| "epoch": 0.582, |
| "grad_norm": 0.72265625, |
| "learning_rate": 8.920006442939772e-05, |
| "loss": 2.7243, |
| "step": 582 |
| }, |
| { |
| "epoch": 0.583, |
| "grad_norm": 0.828125, |
| "learning_rate": 8.885310677936746e-05, |
| "loss": 2.8055, |
| "step": 583 |
| }, |
| { |
| "epoch": 0.584, |
| "grad_norm": 0.66796875, |
| "learning_rate": 8.850628495071336e-05, |
| "loss": 2.8449, |
| "step": 584 |
| }, |
| { |
| "epoch": 0.585, |
| "grad_norm": 0.83203125, |
| "learning_rate": 8.81596031693499e-05, |
| "loss": 2.7153, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.586, |
| "grad_norm": 0.7890625, |
| "learning_rate": 8.781306565948528e-05, |
| "loss": 2.8445, |
| "step": 586 |
| }, |
| { |
| "epoch": 0.587, |
| "grad_norm": 0.60546875, |
| "learning_rate": 8.746667664356956e-05, |
| "loss": 2.7783, |
| "step": 587 |
| }, |
| { |
| "epoch": 0.588, |
| "grad_norm": 0.6953125, |
| "learning_rate": 8.712044034224374e-05, |
| "loss": 2.8369, |
| "step": 588 |
| }, |
| { |
| "epoch": 0.589, |
| "grad_norm": 0.92578125, |
| "learning_rate": 8.677436097428775e-05, |
| "loss": 2.7912, |
| "step": 589 |
| }, |
| { |
| "epoch": 0.59, |
| "grad_norm": 0.6484375, |
| "learning_rate": 8.642844275656957e-05, |
| "loss": 2.8142, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.591, |
| "grad_norm": 0.72265625, |
| "learning_rate": 8.608268990399349e-05, |
| "loss": 2.6924, |
| "step": 591 |
| }, |
| { |
| "epoch": 0.592, |
| "grad_norm": 0.734375, |
| "learning_rate": 8.573710662944885e-05, |
| "loss": 2.7975, |
| "step": 592 |
| }, |
| { |
| "epoch": 0.593, |
| "grad_norm": 0.69140625, |
| "learning_rate": 8.539169714375885e-05, |
| "loss": 2.8273, |
| "step": 593 |
| }, |
| { |
| "epoch": 0.594, |
| "grad_norm": 3.609375, |
| "learning_rate": 8.504646565562906e-05, |
| "loss": 2.7106, |
| "step": 594 |
| }, |
| { |
| "epoch": 0.595, |
| "grad_norm": 0.8359375, |
| "learning_rate": 8.47014163715962e-05, |
| "loss": 2.773, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.596, |
| "grad_norm": 0.8203125, |
| "learning_rate": 8.435655349597689e-05, |
| "loss": 2.759, |
| "step": 596 |
| }, |
| { |
| "epoch": 0.597, |
| "grad_norm": 0.7265625, |
| "learning_rate": 8.401188123081653e-05, |
| "loss": 2.7223, |
| "step": 597 |
| }, |
| { |
| "epoch": 0.598, |
| "grad_norm": 0.73828125, |
| "learning_rate": 8.366740377583781e-05, |
| "loss": 2.7163, |
| "step": 598 |
| }, |
| { |
| "epoch": 0.599, |
| "grad_norm": 0.76953125, |
| "learning_rate": 8.332312532838978e-05, |
| "loss": 2.7544, |
| "step": 599 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 0.71484375, |
| "learning_rate": 8.297905008339677e-05, |
| "loss": 2.7424, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.601, |
| "grad_norm": 0.70703125, |
| "learning_rate": 8.263518223330697e-05, |
| "loss": 2.8959, |
| "step": 601 |
| }, |
| { |
| "epoch": 0.602, |
| "grad_norm": 0.6640625, |
| "learning_rate": 8.229152596804168e-05, |
| "loss": 2.665, |
| "step": 602 |
| }, |
| { |
| "epoch": 0.603, |
| "grad_norm": 0.6640625, |
| "learning_rate": 8.194808547494401e-05, |
| "loss": 2.8718, |
| "step": 603 |
| }, |
| { |
| "epoch": 0.604, |
| "grad_norm": 0.66796875, |
| "learning_rate": 8.160486493872798e-05, |
| "loss": 2.7585, |
| "step": 604 |
| }, |
| { |
| "epoch": 0.605, |
| "grad_norm": 0.65625, |
| "learning_rate": 8.126186854142752e-05, |
| "loss": 2.7911, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.606, |
| "grad_norm": 0.92578125, |
| "learning_rate": 8.091910046234552e-05, |
| "loss": 2.6548, |
| "step": 606 |
| }, |
| { |
| "epoch": 0.607, |
| "grad_norm": 0.625, |
| "learning_rate": 8.057656487800282e-05, |
| "loss": 2.7227, |
| "step": 607 |
| }, |
| { |
| "epoch": 0.608, |
| "grad_norm": 0.66796875, |
| "learning_rate": 8.023426596208739e-05, |
| "loss": 2.7485, |
| "step": 608 |
| }, |
| { |
| "epoch": 0.609, |
| "grad_norm": 0.65625, |
| "learning_rate": 7.989220788540355e-05, |
| "loss": 2.7348, |
| "step": 609 |
| }, |
| { |
| "epoch": 0.61, |
| "grad_norm": 1.546875, |
| "learning_rate": 7.955039481582097e-05, |
| "loss": 2.7158, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.611, |
| "grad_norm": 0.69921875, |
| "learning_rate": 7.920883091822408e-05, |
| "loss": 2.6516, |
| "step": 611 |
| }, |
| { |
| "epoch": 0.612, |
| "grad_norm": 0.6484375, |
| "learning_rate": 7.886752035446114e-05, |
| "loss": 2.8051, |
| "step": 612 |
| }, |
| { |
| "epoch": 0.613, |
| "grad_norm": 0.59765625, |
| "learning_rate": 7.852646728329368e-05, |
| "loss": 2.7428, |
| "step": 613 |
| }, |
| { |
| "epoch": 0.614, |
| "grad_norm": 0.70703125, |
| "learning_rate": 7.818567586034577e-05, |
| "loss": 2.8212, |
| "step": 614 |
| }, |
| { |
| "epoch": 0.615, |
| "grad_norm": 0.66015625, |
| "learning_rate": 7.784515023805328e-05, |
| "loss": 2.7348, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.616, |
| "grad_norm": 0.69921875, |
| "learning_rate": 7.750489456561352e-05, |
| "loss": 2.7394, |
| "step": 616 |
| }, |
| { |
| "epoch": 0.617, |
| "grad_norm": 0.67578125, |
| "learning_rate": 7.716491298893442e-05, |
| "loss": 2.6866, |
| "step": 617 |
| }, |
| { |
| "epoch": 0.618, |
| "grad_norm": 0.67578125, |
| "learning_rate": 7.682520965058428e-05, |
| "loss": 2.6922, |
| "step": 618 |
| }, |
| { |
| "epoch": 0.619, |
| "grad_norm": 0.64453125, |
| "learning_rate": 7.6485788689741e-05, |
| "loss": 2.6569, |
| "step": 619 |
| }, |
| { |
| "epoch": 0.62, |
| "grad_norm": 0.640625, |
| "learning_rate": 7.614665424214193e-05, |
| "loss": 2.752, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.621, |
| "grad_norm": 0.69921875, |
| "learning_rate": 7.580781044003324e-05, |
| "loss": 2.7281, |
| "step": 621 |
| }, |
| { |
| "epoch": 0.622, |
| "grad_norm": 1.171875, |
| "learning_rate": 7.546926141211974e-05, |
| "loss": 2.8616, |
| "step": 622 |
| }, |
| { |
| "epoch": 0.623, |
| "grad_norm": 0.6796875, |
| "learning_rate": 7.513101128351454e-05, |
| "loss": 2.8221, |
| "step": 623 |
| }, |
| { |
| "epoch": 0.624, |
| "grad_norm": 0.640625, |
| "learning_rate": 7.479306417568864e-05, |
| "loss": 2.8145, |
| "step": 624 |
| }, |
| { |
| "epoch": 0.625, |
| "grad_norm": 1.8359375, |
| "learning_rate": 7.445542420642097e-05, |
| "loss": 2.6816, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.626, |
| "grad_norm": 0.640625, |
| "learning_rate": 7.411809548974792e-05, |
| "loss": 2.6947, |
| "step": 626 |
| }, |
| { |
| "epoch": 0.627, |
| "grad_norm": 0.66796875, |
| "learning_rate": 7.378108213591355e-05, |
| "loss": 2.7485, |
| "step": 627 |
| }, |
| { |
| "epoch": 0.628, |
| "grad_norm": 0.66015625, |
| "learning_rate": 7.344438825131911e-05, |
| "loss": 2.7634, |
| "step": 628 |
| }, |
| { |
| "epoch": 0.629, |
| "grad_norm": 0.60546875, |
| "learning_rate": 7.310801793847344e-05, |
| "loss": 2.7374, |
| "step": 629 |
| }, |
| { |
| "epoch": 0.63, |
| "grad_norm": 0.65625, |
| "learning_rate": 7.277197529594257e-05, |
| "loss": 2.6431, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.631, |
| "grad_norm": 0.66796875, |
| "learning_rate": 7.243626441830009e-05, |
| "loss": 2.6069, |
| "step": 631 |
| }, |
| { |
| "epoch": 0.632, |
| "grad_norm": 0.6328125, |
| "learning_rate": 7.210088939607708e-05, |
| "loss": 2.7996, |
| "step": 632 |
| }, |
| { |
| "epoch": 0.633, |
| "grad_norm": 1.046875, |
| "learning_rate": 7.176585431571235e-05, |
| "loss": 2.7437, |
| "step": 633 |
| }, |
| { |
| "epoch": 0.634, |
| "grad_norm": 0.62890625, |
| "learning_rate": 7.143116325950265e-05, |
| "loss": 2.7624, |
| "step": 634 |
| }, |
| { |
| "epoch": 0.635, |
| "grad_norm": 0.59765625, |
| "learning_rate": 7.109682030555283e-05, |
| "loss": 2.715, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.636, |
| "grad_norm": 0.70703125, |
| "learning_rate": 7.076282952772633e-05, |
| "loss": 2.7208, |
| "step": 636 |
| }, |
| { |
| "epoch": 0.637, |
| "grad_norm": 0.6328125, |
| "learning_rate": 7.042919499559537e-05, |
| "loss": 2.7576, |
| "step": 637 |
| }, |
| { |
| "epoch": 0.638, |
| "grad_norm": 0.58203125, |
| "learning_rate": 7.009592077439134e-05, |
| "loss": 2.8255, |
| "step": 638 |
| }, |
| { |
| "epoch": 0.639, |
| "grad_norm": 0.640625, |
| "learning_rate": 6.976301092495556e-05, |
| "loss": 2.7823, |
| "step": 639 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 0.64453125, |
| "learning_rate": 6.943046950368944e-05, |
| "loss": 2.6574, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.641, |
| "grad_norm": 1.2578125, |
| "learning_rate": 6.909830056250527e-05, |
| "loss": 2.6497, |
| "step": 641 |
| }, |
| { |
| "epoch": 0.642, |
| "grad_norm": 0.6171875, |
| "learning_rate": 6.876650814877674e-05, |
| "loss": 2.7926, |
| "step": 642 |
| }, |
| { |
| "epoch": 0.643, |
| "grad_norm": 0.63671875, |
| "learning_rate": 6.843509630528977e-05, |
| "loss": 2.6997, |
| "step": 643 |
| }, |
| { |
| "epoch": 0.644, |
| "grad_norm": 0.61328125, |
| "learning_rate": 6.8104069070193e-05, |
| "loss": 2.7251, |
| "step": 644 |
| }, |
| { |
| "epoch": 0.645, |
| "grad_norm": 0.62109375, |
| "learning_rate": 6.77734304769489e-05, |
| "loss": 2.6186, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.646, |
| "grad_norm": 0.6171875, |
| "learning_rate": 6.744318455428436e-05, |
| "loss": 2.7101, |
| "step": 646 |
| }, |
| { |
| "epoch": 0.647, |
| "grad_norm": 0.66015625, |
| "learning_rate": 6.711333532614168e-05, |
| "loss": 2.6802, |
| "step": 647 |
| }, |
| { |
| "epoch": 0.648, |
| "grad_norm": 0.671875, |
| "learning_rate": 6.67838868116297e-05, |
| "loss": 2.7727, |
| "step": 648 |
| }, |
| { |
| "epoch": 0.649, |
| "grad_norm": 0.640625, |
| "learning_rate": 6.64548430249745e-05, |
| "loss": 2.8216, |
| "step": 649 |
| }, |
| { |
| "epoch": 0.65, |
| "grad_norm": 0.6640625, |
| "learning_rate": 6.612620797547087e-05, |
| "loss": 2.8565, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.651, |
| "grad_norm": 0.671875, |
| "learning_rate": 6.579798566743314e-05, |
| "loss": 2.795, |
| "step": 651 |
| }, |
| { |
| "epoch": 0.652, |
| "grad_norm": 0.66015625, |
| "learning_rate": 6.547018010014654e-05, |
| "loss": 2.7028, |
| "step": 652 |
| }, |
| { |
| "epoch": 0.653, |
| "grad_norm": 0.625, |
| "learning_rate": 6.51427952678185e-05, |
| "loss": 2.7782, |
| "step": 653 |
| }, |
| { |
| "epoch": 0.654, |
| "grad_norm": 0.60546875, |
| "learning_rate": 6.481583515952983e-05, |
| "loss": 2.7242, |
| "step": 654 |
| }, |
| { |
| "epoch": 0.655, |
| "grad_norm": 0.6640625, |
| "learning_rate": 6.448930375918631e-05, |
| "loss": 2.7598, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.656, |
| "grad_norm": 0.64453125, |
| "learning_rate": 6.416320504546997e-05, |
| "loss": 2.6908, |
| "step": 656 |
| }, |
| { |
| "epoch": 0.657, |
| "grad_norm": 1.265625, |
| "learning_rate": 6.383754299179079e-05, |
| "loss": 2.7308, |
| "step": 657 |
| }, |
| { |
| "epoch": 0.658, |
| "grad_norm": 0.66796875, |
| "learning_rate": 6.351232156623803e-05, |
| "loss": 2.7679, |
| "step": 658 |
| }, |
| { |
| "epoch": 0.659, |
| "grad_norm": 0.66015625, |
| "learning_rate": 6.318754473153221e-05, |
| "loss": 2.8747, |
| "step": 659 |
| }, |
| { |
| "epoch": 0.66, |
| "grad_norm": 0.62890625, |
| "learning_rate": 6.286321644497655e-05, |
| "loss": 2.7716, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.661, |
| "grad_norm": 0.62890625, |
| "learning_rate": 6.25393406584088e-05, |
| "loss": 2.6582, |
| "step": 661 |
| }, |
| { |
| "epoch": 0.662, |
| "grad_norm": 0.60546875, |
| "learning_rate": 6.22159213181533e-05, |
| "loss": 2.7907, |
| "step": 662 |
| }, |
| { |
| "epoch": 0.663, |
| "grad_norm": 0.67578125, |
| "learning_rate": 6.18929623649726e-05, |
| "loss": 2.6852, |
| "step": 663 |
| }, |
| { |
| "epoch": 0.664, |
| "grad_norm": 0.61328125, |
| "learning_rate": 6.157046773401964e-05, |
| "loss": 2.7763, |
| "step": 664 |
| }, |
| { |
| "epoch": 0.665, |
| "grad_norm": 0.640625, |
| "learning_rate": 6.12484413547897e-05, |
| "loss": 2.713, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.666, |
| "grad_norm": 0.75390625, |
| "learning_rate": 6.092688715107264e-05, |
| "loss": 2.8589, |
| "step": 666 |
| }, |
| { |
| "epoch": 0.667, |
| "grad_norm": 0.67578125, |
| "learning_rate": 6.0605809040904894e-05, |
| "loss": 2.813, |
| "step": 667 |
| }, |
| { |
| "epoch": 0.668, |
| "grad_norm": 0.68359375, |
| "learning_rate": 6.0285210936521955e-05, |
| "loss": 2.7389, |
| "step": 668 |
| }, |
| { |
| "epoch": 0.669, |
| "grad_norm": 0.625, |
| "learning_rate": 5.9965096744310526e-05, |
| "loss": 2.7049, |
| "step": 669 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.6875, |
| "learning_rate": 5.964547036476099e-05, |
| "loss": 2.7361, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.671, |
| "grad_norm": 0.61328125, |
| "learning_rate": 5.9326335692419995e-05, |
| "loss": 2.7189, |
| "step": 671 |
| }, |
| { |
| "epoch": 0.672, |
| "grad_norm": 1.5078125, |
| "learning_rate": 5.900769661584272e-05, |
| "loss": 2.7206, |
| "step": 672 |
| }, |
| { |
| "epoch": 0.673, |
| "grad_norm": 0.69921875, |
| "learning_rate": 5.868955701754584e-05, |
| "loss": 2.7581, |
| "step": 673 |
| }, |
| { |
| "epoch": 0.674, |
| "grad_norm": 0.63671875, |
| "learning_rate": 5.83719207739599e-05, |
| "loss": 2.5903, |
| "step": 674 |
| }, |
| { |
| "epoch": 0.675, |
| "grad_norm": 0.65234375, |
| "learning_rate": 5.805479175538229e-05, |
| "loss": 2.7786, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.676, |
| "grad_norm": 0.59765625, |
| "learning_rate": 5.773817382593008e-05, |
| "loss": 2.7335, |
| "step": 676 |
| }, |
| { |
| "epoch": 0.677, |
| "grad_norm": 0.6171875, |
| "learning_rate": 5.7422070843492734e-05, |
| "loss": 2.6639, |
| "step": 677 |
| }, |
| { |
| "epoch": 0.678, |
| "grad_norm": 0.66015625, |
| "learning_rate": 5.710648665968543e-05, |
| "loss": 2.6919, |
| "step": 678 |
| }, |
| { |
| "epoch": 0.679, |
| "grad_norm": 0.6328125, |
| "learning_rate": 5.679142511980175e-05, |
| "loss": 2.7755, |
| "step": 679 |
| }, |
| { |
| "epoch": 0.68, |
| "grad_norm": 0.6171875, |
| "learning_rate": 5.647689006276726e-05, |
| "loss": 2.8639, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.681, |
| "grad_norm": 0.6953125, |
| "learning_rate": 5.616288532109225e-05, |
| "loss": 2.9127, |
| "step": 681 |
| }, |
| { |
| "epoch": 0.682, |
| "grad_norm": 0.73046875, |
| "learning_rate": 5.584941472082549e-05, |
| "loss": 2.8527, |
| "step": 682 |
| }, |
| { |
| "epoch": 0.683, |
| "grad_norm": 0.6875, |
| "learning_rate": 5.553648208150728e-05, |
| "loss": 2.7896, |
| "step": 683 |
| }, |
| { |
| "epoch": 0.684, |
| "grad_norm": 0.6171875, |
| "learning_rate": 5.522409121612304e-05, |
| "loss": 2.7857, |
| "step": 684 |
| }, |
| { |
| "epoch": 0.685, |
| "grad_norm": 0.65625, |
| "learning_rate": 5.491224593105695e-05, |
| "loss": 2.7679, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.686, |
| "grad_norm": 0.6640625, |
| "learning_rate": 5.4600950026045326e-05, |
| "loss": 2.7092, |
| "step": 686 |
| }, |
| { |
| "epoch": 0.687, |
| "grad_norm": 0.64453125, |
| "learning_rate": 5.4290207294130615e-05, |
| "loss": 2.7738, |
| "step": 687 |
| }, |
| { |
| "epoch": 0.688, |
| "grad_norm": 0.99609375, |
| "learning_rate": 5.398002152161484e-05, |
| "loss": 2.7504, |
| "step": 688 |
| }, |
| { |
| "epoch": 0.689, |
| "grad_norm": 0.609375, |
| "learning_rate": 5.3670396488013854e-05, |
| "loss": 2.783, |
| "step": 689 |
| }, |
| { |
| "epoch": 0.69, |
| "grad_norm": 0.640625, |
| "learning_rate": 5.33613359660109e-05, |
| "loss": 2.7702, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.691, |
| "grad_norm": 0.66796875, |
| "learning_rate": 5.305284372141095e-05, |
| "loss": 2.7703, |
| "step": 691 |
| }, |
| { |
| "epoch": 0.692, |
| "grad_norm": 0.67578125, |
| "learning_rate": 5.274492351309461e-05, |
| "loss": 2.8481, |
| "step": 692 |
| }, |
| { |
| "epoch": 0.693, |
| "grad_norm": 0.640625, |
| "learning_rate": 5.243757909297247e-05, |
| "loss": 2.8131, |
| "step": 693 |
| }, |
| { |
| "epoch": 0.694, |
| "grad_norm": 0.62109375, |
| "learning_rate": 5.213081420593933e-05, |
| "loss": 2.8698, |
| "step": 694 |
| }, |
| { |
| "epoch": 0.695, |
| "grad_norm": 0.68359375, |
| "learning_rate": 5.182463258982846e-05, |
| "loss": 2.8107, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.696, |
| "grad_norm": 0.6796875, |
| "learning_rate": 5.15190379753663e-05, |
| "loss": 2.8036, |
| "step": 696 |
| }, |
| { |
| "epoch": 0.697, |
| "grad_norm": 0.64453125, |
| "learning_rate": 5.121403408612672e-05, |
| "loss": 2.8471, |
| "step": 697 |
| }, |
| { |
| "epoch": 0.698, |
| "grad_norm": 0.609375, |
| "learning_rate": 5.090962463848592e-05, |
| "loss": 2.8396, |
| "step": 698 |
| }, |
| { |
| "epoch": 0.699, |
| "grad_norm": 0.62890625, |
| "learning_rate": 5.0605813341576924e-05, |
| "loss": 2.8636, |
| "step": 699 |
| }, |
| { |
| "epoch": 0.7, |
| "grad_norm": 0.70703125, |
| "learning_rate": 5.0302603897244474e-05, |
| "loss": 2.9405, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.701, |
| "grad_norm": 0.62890625, |
| "learning_rate": 5.000000000000002e-05, |
| "loss": 2.7732, |
| "step": 701 |
| }, |
| { |
| "epoch": 0.702, |
| "grad_norm": 0.671875, |
| "learning_rate": 4.969800533697649e-05, |
| "loss": 2.9202, |
| "step": 702 |
| }, |
| { |
| "epoch": 0.703, |
| "grad_norm": 0.65234375, |
| "learning_rate": 4.939662358788364e-05, |
| "loss": 2.8386, |
| "step": 703 |
| }, |
| { |
| "epoch": 0.704, |
| "grad_norm": 0.82421875, |
| "learning_rate": 4.909585842496287e-05, |
| "loss": 2.8088, |
| "step": 704 |
| }, |
| { |
| "epoch": 0.705, |
| "grad_norm": 0.66796875, |
| "learning_rate": 4.8795713512942865e-05, |
| "loss": 2.6467, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.706, |
| "grad_norm": 0.65234375, |
| "learning_rate": 4.8496192508994576e-05, |
| "loss": 2.7861, |
| "step": 706 |
| }, |
| { |
| "epoch": 0.707, |
| "grad_norm": 0.52734375, |
| "learning_rate": 4.8197299062686995e-05, |
| "loss": 2.7869, |
| "step": 707 |
| }, |
| { |
| "epoch": 0.708, |
| "grad_norm": 0.6015625, |
| "learning_rate": 4.78990368159424e-05, |
| "loss": 2.6457, |
| "step": 708 |
| }, |
| { |
| "epoch": 0.709, |
| "grad_norm": 0.57421875, |
| "learning_rate": 4.7601409402992106e-05, |
| "loss": 2.6031, |
| "step": 709 |
| }, |
| { |
| "epoch": 0.71, |
| "grad_norm": 0.58203125, |
| "learning_rate": 4.7304420450332244e-05, |
| "loss": 2.8342, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.711, |
| "grad_norm": 0.56640625, |
| "learning_rate": 4.700807357667952e-05, |
| "loss": 2.7622, |
| "step": 711 |
| }, |
| { |
| "epoch": 0.712, |
| "grad_norm": 0.61328125, |
| "learning_rate": 4.6712372392927e-05, |
| "loss": 2.7514, |
| "step": 712 |
| }, |
| { |
| "epoch": 0.713, |
| "grad_norm": 0.6640625, |
| "learning_rate": 4.6417320502100316e-05, |
| "loss": 2.8732, |
| "step": 713 |
| }, |
| { |
| "epoch": 0.714, |
| "grad_norm": 0.6015625, |
| "learning_rate": 4.612292149931369e-05, |
| "loss": 2.8161, |
| "step": 714 |
| }, |
| { |
| "epoch": 0.715, |
| "grad_norm": 0.578125, |
| "learning_rate": 4.582917897172603e-05, |
| "loss": 2.7898, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.716, |
| "grad_norm": 0.59765625, |
| "learning_rate": 4.5536096498497295e-05, |
| "loss": 2.7848, |
| "step": 716 |
| }, |
| { |
| "epoch": 0.717, |
| "grad_norm": 0.59765625, |
| "learning_rate": 4.524367765074499e-05, |
| "loss": 2.703, |
| "step": 717 |
| }, |
| { |
| "epoch": 0.718, |
| "grad_norm": 0.59375, |
| "learning_rate": 4.495192599150044e-05, |
| "loss": 2.7208, |
| "step": 718 |
| }, |
| { |
| "epoch": 0.719, |
| "grad_norm": 1.3671875, |
| "learning_rate": 4.46608450756656e-05, |
| "loss": 2.767, |
| "step": 719 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 0.57421875, |
| "learning_rate": 4.437043844996952e-05, |
| "loss": 2.7353, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.721, |
| "grad_norm": 0.640625, |
| "learning_rate": 4.4080709652925336e-05, |
| "loss": 2.7186, |
| "step": 721 |
| }, |
| { |
| "epoch": 0.722, |
| "grad_norm": 18.625, |
| "learning_rate": 4.379166221478697e-05, |
| "loss": 2.6783, |
| "step": 722 |
| }, |
| { |
| "epoch": 0.723, |
| "grad_norm": 0.6015625, |
| "learning_rate": 4.350329965750621e-05, |
| "loss": 2.718, |
| "step": 723 |
| }, |
| { |
| "epoch": 0.724, |
| "grad_norm": 0.64453125, |
| "learning_rate": 4.32156254946899e-05, |
| "loss": 2.6782, |
| "step": 724 |
| }, |
| { |
| "epoch": 0.725, |
| "grad_norm": 0.6484375, |
| "learning_rate": 4.2928643231556844e-05, |
| "loss": 2.7787, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.726, |
| "grad_norm": 0.59765625, |
| "learning_rate": 4.264235636489542e-05, |
| "loss": 2.7355, |
| "step": 726 |
| }, |
| { |
| "epoch": 0.727, |
| "grad_norm": 0.59765625, |
| "learning_rate": 4.235676838302068e-05, |
| "loss": 2.8365, |
| "step": 727 |
| }, |
| { |
| "epoch": 0.728, |
| "grad_norm": 0.59765625, |
| "learning_rate": 4.207188276573214e-05, |
| "loss": 2.7514, |
| "step": 728 |
| }, |
| { |
| "epoch": 0.729, |
| "grad_norm": 0.60546875, |
| "learning_rate": 4.1787702984271074e-05, |
| "loss": 2.7733, |
| "step": 729 |
| }, |
| { |
| "epoch": 0.73, |
| "grad_norm": 1.8984375, |
| "learning_rate": 4.150423250127845e-05, |
| "loss": 2.809, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.731, |
| "grad_norm": 0.5859375, |
| "learning_rate": 4.12214747707527e-05, |
| "loss": 2.7394, |
| "step": 731 |
| }, |
| { |
| "epoch": 0.732, |
| "grad_norm": 0.578125, |
| "learning_rate": 4.093943323800745e-05, |
| "loss": 2.6793, |
| "step": 732 |
| }, |
| { |
| "epoch": 0.733, |
| "grad_norm": 0.5625, |
| "learning_rate": 4.065811133962987e-05, |
| "loss": 2.7069, |
| "step": 733 |
| }, |
| { |
| "epoch": 0.734, |
| "grad_norm": 0.57421875, |
| "learning_rate": 4.037751250343841e-05, |
| "loss": 2.6702, |
| "step": 734 |
| }, |
| { |
| "epoch": 0.735, |
| "grad_norm": 1.109375, |
| "learning_rate": 4.009764014844143e-05, |
| "loss": 2.8871, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.736, |
| "grad_norm": 0.62890625, |
| "learning_rate": 3.981849768479517e-05, |
| "loss": 2.7682, |
| "step": 736 |
| }, |
| { |
| "epoch": 0.737, |
| "grad_norm": 0.59765625, |
| "learning_rate": 3.954008851376252e-05, |
| "loss": 2.7024, |
| "step": 737 |
| }, |
| { |
| "epoch": 0.738, |
| "grad_norm": 0.59765625, |
| "learning_rate": 3.9262416027671356e-05, |
| "loss": 2.6956, |
| "step": 738 |
| }, |
| { |
| "epoch": 0.739, |
| "grad_norm": 0.6640625, |
| "learning_rate": 3.8985483609873244e-05, |
| "loss": 2.8372, |
| "step": 739 |
| }, |
| { |
| "epoch": 0.74, |
| "grad_norm": 0.5703125, |
| "learning_rate": 3.8709294634702376e-05, |
| "loss": 2.7524, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.741, |
| "grad_norm": 0.59375, |
| "learning_rate": 3.843385246743417e-05, |
| "loss": 2.8114, |
| "step": 741 |
| }, |
| { |
| "epoch": 0.742, |
| "grad_norm": 0.60546875, |
| "learning_rate": 3.8159160464244606e-05, |
| "loss": 2.8749, |
| "step": 742 |
| }, |
| { |
| "epoch": 0.743, |
| "grad_norm": 0.578125, |
| "learning_rate": 3.788522197216897e-05, |
| "loss": 2.7037, |
| "step": 743 |
| }, |
| { |
| "epoch": 0.744, |
| "grad_norm": 0.55078125, |
| "learning_rate": 3.7612040329061405e-05, |
| "loss": 2.7504, |
| "step": 744 |
| }, |
| { |
| "epoch": 0.745, |
| "grad_norm": 0.59765625, |
| "learning_rate": 3.733961886355398e-05, |
| "loss": 2.7219, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.746, |
| "grad_norm": 0.60546875, |
| "learning_rate": 3.7067960895016275e-05, |
| "loss": 2.741, |
| "step": 746 |
| }, |
| { |
| "epoch": 0.747, |
| "grad_norm": 0.5625, |
| "learning_rate": 3.679706973351491e-05, |
| "loss": 2.7085, |
| "step": 747 |
| }, |
| { |
| "epoch": 0.748, |
| "grad_norm": 0.59375, |
| "learning_rate": 3.6526948679773257e-05, |
| "loss": 2.7705, |
| "step": 748 |
| }, |
| { |
| "epoch": 0.749, |
| "grad_norm": 0.58984375, |
| "learning_rate": 3.6257601025131026e-05, |
| "loss": 2.7403, |
| "step": 749 |
| }, |
| { |
| "epoch": 0.75, |
| "grad_norm": 0.9375, |
| "learning_rate": 3.5989030051504434e-05, |
| "loss": 2.7612, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.75, |
| "eval_loss": 2.7113237380981445, |
| "eval_runtime": 31.7342, |
| "eval_samples_per_second": 15.756, |
| "eval_steps_per_second": 2.647, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.751, |
| "grad_norm": 0.59375, |
| "learning_rate": 3.5721239031346066e-05, |
| "loss": 2.8165, |
| "step": 751 |
| }, |
| { |
| "epoch": 0.752, |
| "grad_norm": 0.5625, |
| "learning_rate": 3.545423122760493e-05, |
| "loss": 2.6717, |
| "step": 752 |
| }, |
| { |
| "epoch": 0.753, |
| "grad_norm": 0.578125, |
| "learning_rate": 3.518800989368691e-05, |
| "loss": 2.5895, |
| "step": 753 |
| }, |
| { |
| "epoch": 0.754, |
| "grad_norm": 0.5625, |
| "learning_rate": 3.492257827341492e-05, |
| "loss": 2.7452, |
| "step": 754 |
| }, |
| { |
| "epoch": 0.755, |
| "grad_norm": 0.5625, |
| "learning_rate": 3.465793960098945e-05, |
| "loss": 2.7445, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.756, |
| "grad_norm": 0.59765625, |
| "learning_rate": 3.439409710094929e-05, |
| "loss": 2.805, |
| "step": 756 |
| }, |
| { |
| "epoch": 0.757, |
| "grad_norm": 0.640625, |
| "learning_rate": 3.413105398813195e-05, |
| "loss": 2.8297, |
| "step": 757 |
| }, |
| { |
| "epoch": 0.758, |
| "grad_norm": 0.55078125, |
| "learning_rate": 3.386881346763483e-05, |
| "loss": 2.6497, |
| "step": 758 |
| }, |
| { |
| "epoch": 0.759, |
| "grad_norm": 0.56640625, |
| "learning_rate": 3.360737873477584e-05, |
| "loss": 2.7854, |
| "step": 759 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.58203125, |
| "learning_rate": 3.334675297505476e-05, |
| "loss": 2.7295, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.761, |
| "grad_norm": 0.578125, |
| "learning_rate": 3.308693936411421e-05, |
| "loss": 2.7563, |
| "step": 761 |
| }, |
| { |
| "epoch": 0.762, |
| "grad_norm": 0.5859375, |
| "learning_rate": 3.2827941067700996e-05, |
| "loss": 2.8956, |
| "step": 762 |
| }, |
| { |
| "epoch": 0.763, |
| "grad_norm": 0.76953125, |
| "learning_rate": 3.2569761241627696e-05, |
| "loss": 2.7614, |
| "step": 763 |
| }, |
| { |
| "epoch": 0.764, |
| "grad_norm": 0.578125, |
| "learning_rate": 3.231240303173394e-05, |
| "loss": 2.7303, |
| "step": 764 |
| }, |
| { |
| "epoch": 0.765, |
| "grad_norm": 0.58984375, |
| "learning_rate": 3.205586957384838e-05, |
| "loss": 2.7826, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.766, |
| "grad_norm": 1.3359375, |
| "learning_rate": 3.1800163993750166e-05, |
| "loss": 2.7078, |
| "step": 766 |
| }, |
| { |
| "epoch": 0.767, |
| "grad_norm": 0.54296875, |
| "learning_rate": 3.154528940713113e-05, |
| "loss": 2.6754, |
| "step": 767 |
| }, |
| { |
| "epoch": 0.768, |
| "grad_norm": 0.59765625, |
| "learning_rate": 3.129124891955771e-05, |
| "loss": 2.7744, |
| "step": 768 |
| }, |
| { |
| "epoch": 0.769, |
| "grad_norm": 0.5625, |
| "learning_rate": 3.103804562643302e-05, |
| "loss": 2.653, |
| "step": 769 |
| }, |
| { |
| "epoch": 0.77, |
| "grad_norm": 0.546875, |
| "learning_rate": 3.078568261295933e-05, |
| "loss": 2.7278, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.771, |
| "grad_norm": 0.57421875, |
| "learning_rate": 3.053416295410026e-05, |
| "loss": 2.7533, |
| "step": 771 |
| }, |
| { |
| "epoch": 0.772, |
| "grad_norm": 0.60546875, |
| "learning_rate": 3.0283489714543556e-05, |
| "loss": 2.7639, |
| "step": 772 |
| }, |
| { |
| "epoch": 0.773, |
| "grad_norm": 0.578125, |
| "learning_rate": 3.0033665948663448e-05, |
| "loss": 2.7951, |
| "step": 773 |
| }, |
| { |
| "epoch": 0.774, |
| "grad_norm": 0.6015625, |
| "learning_rate": 2.9784694700483762e-05, |
| "loss": 2.7375, |
| "step": 774 |
| }, |
| { |
| "epoch": 0.775, |
| "grad_norm": 0.55859375, |
| "learning_rate": 2.953657900364053e-05, |
| "loss": 2.7481, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.776, |
| "grad_norm": 0.59765625, |
| "learning_rate": 2.9289321881345254e-05, |
| "loss": 2.7437, |
| "step": 776 |
| }, |
| { |
| "epoch": 0.777, |
| "grad_norm": 0.58984375, |
| "learning_rate": 2.904292634634793e-05, |
| "loss": 2.7637, |
| "step": 777 |
| }, |
| { |
| "epoch": 0.778, |
| "grad_norm": 0.546875, |
| "learning_rate": 2.879739540090036e-05, |
| "loss": 2.7252, |
| "step": 778 |
| }, |
| { |
| "epoch": 0.779, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.8552732036719687e-05, |
| "loss": 2.5999, |
| "step": 779 |
| }, |
| { |
| "epoch": 0.78, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.8308939234951726e-05, |
| "loss": 2.8094, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.781, |
| "grad_norm": 0.58984375, |
| "learning_rate": 2.8066019966134904e-05, |
| "loss": 2.7139, |
| "step": 781 |
| }, |
| { |
| "epoch": 0.782, |
| "grad_norm": 1.46875, |
| "learning_rate": 2.7823977190163786e-05, |
| "loss": 2.8623, |
| "step": 782 |
| }, |
| { |
| "epoch": 0.783, |
| "grad_norm": 0.52734375, |
| "learning_rate": 2.7582813856253275e-05, |
| "loss": 2.7122, |
| "step": 783 |
| }, |
| { |
| "epoch": 0.784, |
| "grad_norm": 0.55078125, |
| "learning_rate": 2.734253290290242e-05, |
| "loss": 2.7772, |
| "step": 784 |
| }, |
| { |
| "epoch": 0.785, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.7103137257858868e-05, |
| "loss": 2.7612, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.786, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.6864629838082956e-05, |
| "loss": 2.7303, |
| "step": 786 |
| }, |
| { |
| "epoch": 0.787, |
| "grad_norm": 0.546875, |
| "learning_rate": 2.6627013549712355e-05, |
| "loss": 2.6967, |
| "step": 787 |
| }, |
| { |
| "epoch": 0.788, |
| "grad_norm": 0.5234375, |
| "learning_rate": 2.639029128802657e-05, |
| "loss": 2.675, |
| "step": 788 |
| }, |
| { |
| "epoch": 0.789, |
| "grad_norm": 0.52734375, |
| "learning_rate": 2.615446593741161e-05, |
| "loss": 2.771, |
| "step": 789 |
| }, |
| { |
| "epoch": 0.79, |
| "grad_norm": 0.59765625, |
| "learning_rate": 2.5919540371325e-05, |
| "loss": 2.7169, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.791, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.5685517452260567e-05, |
| "loss": 2.6604, |
| "step": 791 |
| }, |
| { |
| "epoch": 0.792, |
| "grad_norm": 0.59765625, |
| "learning_rate": 2.5452400031713785e-05, |
| "loss": 2.6749, |
| "step": 792 |
| }, |
| { |
| "epoch": 0.793, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.5220190950146827e-05, |
| "loss": 2.6759, |
| "step": 793 |
| }, |
| { |
| "epoch": 0.794, |
| "grad_norm": 0.57421875, |
| "learning_rate": 2.4988893036954043e-05, |
| "loss": 2.5863, |
| "step": 794 |
| }, |
| { |
| "epoch": 0.795, |
| "grad_norm": 0.55078125, |
| "learning_rate": 2.4758509110427575e-05, |
| "loss": 2.7512, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.796, |
| "grad_norm": 0.59765625, |
| "learning_rate": 2.45290419777228e-05, |
| "loss": 2.7619, |
| "step": 796 |
| }, |
| { |
| "epoch": 0.797, |
| "grad_norm": 0.74609375, |
| "learning_rate": 2.4300494434824373e-05, |
| "loss": 2.6975, |
| "step": 797 |
| }, |
| { |
| "epoch": 0.798, |
| "grad_norm": 0.56640625, |
| "learning_rate": 2.407286926651192e-05, |
| "loss": 2.6837, |
| "step": 798 |
| }, |
| { |
| "epoch": 0.799, |
| "grad_norm": 0.5234375, |
| "learning_rate": 2.3846169246326343e-05, |
| "loss": 2.7131, |
| "step": 799 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 0.54296875, |
| "learning_rate": 2.362039713653581e-05, |
| "loss": 2.7429, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.801, |
| "grad_norm": 0.578125, |
| "learning_rate": 2.339555568810221e-05, |
| "loss": 2.6673, |
| "step": 801 |
| }, |
| { |
| "epoch": 0.802, |
| "grad_norm": 0.57421875, |
| "learning_rate": 2.3171647640647687e-05, |
| "loss": 2.6941, |
| "step": 802 |
| }, |
| { |
| "epoch": 0.803, |
| "grad_norm": 0.6484375, |
| "learning_rate": 2.2948675722421086e-05, |
| "loss": 2.735, |
| "step": 803 |
| }, |
| { |
| "epoch": 0.804, |
| "grad_norm": 0.79296875, |
| "learning_rate": 2.2726642650264895e-05, |
| "loss": 2.7529, |
| "step": 804 |
| }, |
| { |
| "epoch": 0.805, |
| "grad_norm": 0.55859375, |
| "learning_rate": 2.2505551129582047e-05, |
| "loss": 2.6125, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.806, |
| "grad_norm": 0.56640625, |
| "learning_rate": 2.2285403854302912e-05, |
| "loss": 2.8059, |
| "step": 806 |
| }, |
| { |
| "epoch": 0.807, |
| "grad_norm": 0.546875, |
| "learning_rate": 2.2066203506852566e-05, |
| "loss": 2.6916, |
| "step": 807 |
| }, |
| { |
| "epoch": 0.808, |
| "grad_norm": 0.7265625, |
| "learning_rate": 2.1847952758118117e-05, |
| "loss": 2.7839, |
| "step": 808 |
| }, |
| { |
| "epoch": 0.809, |
| "grad_norm": 0.58984375, |
| "learning_rate": 2.163065426741603e-05, |
| "loss": 2.7451, |
| "step": 809 |
| }, |
| { |
| "epoch": 0.81, |
| "grad_norm": 0.53515625, |
| "learning_rate": 2.1414310682459802e-05, |
| "loss": 2.7282, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.811, |
| "grad_norm": 0.57421875, |
| "learning_rate": 2.119892463932781e-05, |
| "loss": 2.6989, |
| "step": 811 |
| }, |
| { |
| "epoch": 0.812, |
| "grad_norm": 0.5859375, |
| "learning_rate": 2.098449876243096e-05, |
| "loss": 2.6681, |
| "step": 812 |
| }, |
| { |
| "epoch": 0.813, |
| "grad_norm": 0.70703125, |
| "learning_rate": 2.0771035664480942e-05, |
| "loss": 2.7579, |
| "step": 813 |
| }, |
| { |
| "epoch": 0.814, |
| "grad_norm": 0.57421875, |
| "learning_rate": 2.0558537946458177e-05, |
| "loss": 2.6096, |
| "step": 814 |
| }, |
| { |
| "epoch": 0.815, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.0347008197580374e-05, |
| "loss": 2.5568, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.816, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.013644899527074e-05, |
| "loss": 2.7679, |
| "step": 816 |
| }, |
| { |
| "epoch": 0.817, |
| "grad_norm": 0.5546875, |
| "learning_rate": 1.9926862905126665e-05, |
| "loss": 2.751, |
| "step": 817 |
| }, |
| { |
| "epoch": 0.818, |
| "grad_norm": 0.51953125, |
| "learning_rate": 1.9718252480888566e-05, |
| "loss": 2.7757, |
| "step": 818 |
| }, |
| { |
| "epoch": 0.819, |
| "grad_norm": 0.58203125, |
| "learning_rate": 1.9510620264408596e-05, |
| "loss": 2.6745, |
| "step": 819 |
| }, |
| { |
| "epoch": 0.82, |
| "grad_norm": 0.58203125, |
| "learning_rate": 1.930396878561983e-05, |
| "loss": 2.6728, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.821, |
| "grad_norm": 0.5859375, |
| "learning_rate": 1.9098300562505266e-05, |
| "loss": 2.6893, |
| "step": 821 |
| }, |
| { |
| "epoch": 0.822, |
| "grad_norm": 0.58203125, |
| "learning_rate": 1.8893618101067355e-05, |
| "loss": 2.8697, |
| "step": 822 |
| }, |
| { |
| "epoch": 0.823, |
| "grad_norm": 0.5546875, |
| "learning_rate": 1.8689923895297245e-05, |
| "loss": 2.7268, |
| "step": 823 |
| }, |
| { |
| "epoch": 0.824, |
| "grad_norm": 0.578125, |
| "learning_rate": 1.848722042714457e-05, |
| "loss": 2.7974, |
| "step": 824 |
| }, |
| { |
| "epoch": 0.825, |
| "grad_norm": 0.53515625, |
| "learning_rate": 1.8285510166487152e-05, |
| "loss": 2.691, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.826, |
| "grad_norm": 0.5859375, |
| "learning_rate": 1.808479557110081e-05, |
| "loss": 2.6877, |
| "step": 826 |
| }, |
| { |
| "epoch": 0.827, |
| "grad_norm": 0.53125, |
| "learning_rate": 1.78850790866296e-05, |
| "loss": 2.6971, |
| "step": 827 |
| }, |
| { |
| "epoch": 0.828, |
| "grad_norm": 0.5625, |
| "learning_rate": 1.7686363146555805e-05, |
| "loss": 2.8624, |
| "step": 828 |
| }, |
| { |
| "epoch": 0.829, |
| "grad_norm": 2.421875, |
| "learning_rate": 1.7488650172170496e-05, |
| "loss": 2.6911, |
| "step": 829 |
| }, |
| { |
| "epoch": 0.83, |
| "grad_norm": 0.5859375, |
| "learning_rate": 1.7291942572543807e-05, |
| "loss": 2.6905, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.831, |
| "grad_norm": 0.58984375, |
| "learning_rate": 1.7096242744495837e-05, |
| "loss": 2.7459, |
| "step": 831 |
| }, |
| { |
| "epoch": 0.832, |
| "grad_norm": 0.6875, |
| "learning_rate": 1.690155307256719e-05, |
| "loss": 2.738, |
| "step": 832 |
| }, |
| { |
| "epoch": 0.833, |
| "grad_norm": 0.578125, |
| "learning_rate": 1.6707875928990058e-05, |
| "loss": 2.7459, |
| "step": 833 |
| }, |
| { |
| "epoch": 0.834, |
| "grad_norm": 0.5859375, |
| "learning_rate": 1.6515213673659357e-05, |
| "loss": 2.7432, |
| "step": 834 |
| }, |
| { |
| "epoch": 0.835, |
| "grad_norm": 0.55078125, |
| "learning_rate": 1.632356865410384e-05, |
| "loss": 2.6985, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.836, |
| "grad_norm": 0.58203125, |
| "learning_rate": 1.6132943205457606e-05, |
| "loss": 2.8408, |
| "step": 836 |
| }, |
| { |
| "epoch": 0.837, |
| "grad_norm": 0.53515625, |
| "learning_rate": 1.5943339650431576e-05, |
| "loss": 2.662, |
| "step": 837 |
| }, |
| { |
| "epoch": 0.838, |
| "grad_norm": 0.53515625, |
| "learning_rate": 1.5754760299285252e-05, |
| "loss": 2.85, |
| "step": 838 |
| }, |
| { |
| "epoch": 0.839, |
| "grad_norm": 0.5390625, |
| "learning_rate": 1.5567207449798515e-05, |
| "loss": 2.6535, |
| "step": 839 |
| }, |
| { |
| "epoch": 0.84, |
| "grad_norm": 0.55078125, |
| "learning_rate": 1.538068338724361e-05, |
| "loss": 2.7144, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.841, |
| "grad_norm": 0.53515625, |
| "learning_rate": 1.5195190384357404e-05, |
| "loss": 2.8185, |
| "step": 841 |
| }, |
| { |
| "epoch": 0.842, |
| "grad_norm": 1.0703125, |
| "learning_rate": 1.5010730701313625e-05, |
| "loss": 2.7978, |
| "step": 842 |
| }, |
| { |
| "epoch": 0.843, |
| "grad_norm": 0.546875, |
| "learning_rate": 1.4827306585695234e-05, |
| "loss": 2.7138, |
| "step": 843 |
| }, |
| { |
| "epoch": 0.844, |
| "grad_norm": 0.890625, |
| "learning_rate": 1.4644920272467244e-05, |
| "loss": 2.7772, |
| "step": 844 |
| }, |
| { |
| "epoch": 0.845, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.4463573983949341e-05, |
| "loss": 2.7598, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.846, |
| "grad_norm": 0.6953125, |
| "learning_rate": 1.4283269929788779e-05, |
| "loss": 2.7201, |
| "step": 846 |
| }, |
| { |
| "epoch": 0.847, |
| "grad_norm": 0.51953125, |
| "learning_rate": 1.4104010306933557e-05, |
| "loss": 2.8091, |
| "step": 847 |
| }, |
| { |
| "epoch": 0.848, |
| "grad_norm": 0.5546875, |
| "learning_rate": 1.3925797299605647e-05, |
| "loss": 2.7008, |
| "step": 848 |
| }, |
| { |
| "epoch": 0.849, |
| "grad_norm": 0.52734375, |
| "learning_rate": 1.3748633079274253e-05, |
| "loss": 2.655, |
| "step": 849 |
| }, |
| { |
| "epoch": 0.85, |
| "grad_norm": 0.57421875, |
| "learning_rate": 1.3572519804629536e-05, |
| "loss": 2.6479, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.851, |
| "grad_norm": 0.5625, |
| "learning_rate": 1.339745962155613e-05, |
| "loss": 2.8183, |
| "step": 851 |
| }, |
| { |
| "epoch": 0.852, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.3223454663107172e-05, |
| "loss": 2.7465, |
| "step": 852 |
| }, |
| { |
| "epoch": 0.853, |
| "grad_norm": 0.6015625, |
| "learning_rate": 1.30505070494781e-05, |
| "loss": 2.7232, |
| "step": 853 |
| }, |
| { |
| "epoch": 0.854, |
| "grad_norm": 0.5078125, |
| "learning_rate": 1.2878618887981064e-05, |
| "loss": 2.7987, |
| "step": 854 |
| }, |
| { |
| "epoch": 0.855, |
| "grad_norm": 0.52734375, |
| "learning_rate": 1.2707792273019048e-05, |
| "loss": 2.7005, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.856, |
| "grad_norm": 0.55078125, |
| "learning_rate": 1.2538029286060426e-05, |
| "loss": 2.7156, |
| "step": 856 |
| }, |
| { |
| "epoch": 0.857, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.2369331995613665e-05, |
| "loss": 2.757, |
| "step": 857 |
| }, |
| { |
| "epoch": 0.858, |
| "grad_norm": 0.91015625, |
| "learning_rate": 1.2201702457201947e-05, |
| "loss": 2.6486, |
| "step": 858 |
| }, |
| { |
| "epoch": 0.859, |
| "grad_norm": 0.5703125, |
| "learning_rate": 1.2035142713338366e-05, |
| "loss": 2.7237, |
| "step": 859 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 1.125, |
| "learning_rate": 1.1869654793500784e-05, |
| "loss": 2.6579, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.861, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.1705240714107302e-05, |
| "loss": 2.719, |
| "step": 861 |
| }, |
| { |
| "epoch": 0.862, |
| "grad_norm": 0.51171875, |
| "learning_rate": 1.1541902478491606e-05, |
| "loss": 2.6551, |
| "step": 862 |
| }, |
| { |
| "epoch": 0.863, |
| "grad_norm": 0.52734375, |
| "learning_rate": 1.1379642076878527e-05, |
| "loss": 2.6851, |
| "step": 863 |
| }, |
| { |
| "epoch": 0.864, |
| "grad_norm": 0.5390625, |
| "learning_rate": 1.1218461486359877e-05, |
| "loss": 2.7538, |
| "step": 864 |
| }, |
| { |
| "epoch": 0.865, |
| "grad_norm": 0.58984375, |
| "learning_rate": 1.1058362670870249e-05, |
| "loss": 2.693, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.866, |
| "grad_norm": 0.578125, |
| "learning_rate": 1.0899347581163221e-05, |
| "loss": 2.7243, |
| "step": 866 |
| }, |
| { |
| "epoch": 0.867, |
| "grad_norm": 0.546875, |
| "learning_rate": 1.0741418154787442e-05, |
| "loss": 2.6959, |
| "step": 867 |
| }, |
| { |
| "epoch": 0.868, |
| "grad_norm": 0.5703125, |
| "learning_rate": 1.0584576316063188e-05, |
| "loss": 2.7305, |
| "step": 868 |
| }, |
| { |
| "epoch": 0.869, |
| "grad_norm": 0.57421875, |
| "learning_rate": 1.042882397605871e-05, |
| "loss": 2.7807, |
| "step": 869 |
| }, |
| { |
| "epoch": 0.87, |
| "grad_norm": 0.5546875, |
| "learning_rate": 1.0274163032567163e-05, |
| "loss": 2.7001, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.871, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.0120595370083318e-05, |
| "loss": 2.7309, |
| "step": 871 |
| }, |
| { |
| "epoch": 0.872, |
| "grad_norm": 0.54296875, |
| "learning_rate": 9.968122859780648e-06, |
| "loss": 2.6748, |
| "step": 872 |
| }, |
| { |
| "epoch": 0.873, |
| "grad_norm": 0.58203125, |
| "learning_rate": 9.816747359488632e-06, |
| "loss": 2.7391, |
| "step": 873 |
| }, |
| { |
| "epoch": 0.874, |
| "grad_norm": 0.5625, |
| "learning_rate": 9.666470713669918e-06, |
| "loss": 2.8011, |
| "step": 874 |
| }, |
| { |
| "epoch": 0.875, |
| "grad_norm": 1.21875, |
| "learning_rate": 9.517294753398064e-06, |
| "loss": 2.6658, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.876, |
| "grad_norm": 0.546875, |
| "learning_rate": 9.369221296335006e-06, |
| "loss": 2.7542, |
| "step": 876 |
| }, |
| { |
| "epoch": 0.877, |
| "grad_norm": 0.5390625, |
| "learning_rate": 9.222252146709142e-06, |
| "loss": 2.6932, |
| "step": 877 |
| }, |
| { |
| "epoch": 0.878, |
| "grad_norm": 0.51953125, |
| "learning_rate": 9.076389095293148e-06, |
| "loss": 2.7678, |
| "step": 878 |
| }, |
| { |
| "epoch": 0.879, |
| "grad_norm": 0.53515625, |
| "learning_rate": 8.931633919382298e-06, |
| "loss": 2.6873, |
| "step": 879 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 0.52734375, |
| "learning_rate": 8.787988382772705e-06, |
| "loss": 2.812, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.881, |
| "grad_norm": 0.57421875, |
| "learning_rate": 8.645454235739903e-06, |
| "loss": 2.6715, |
| "step": 881 |
| }, |
| { |
| "epoch": 0.882, |
| "grad_norm": 0.53125, |
| "learning_rate": 8.504033215017527e-06, |
| "loss": 2.7504, |
| "step": 882 |
| }, |
| { |
| "epoch": 0.883, |
| "grad_norm": 0.515625, |
| "learning_rate": 8.363727043776038e-06, |
| "loss": 2.7513, |
| "step": 883 |
| }, |
| { |
| "epoch": 0.884, |
| "grad_norm": 0.51171875, |
| "learning_rate": 8.224537431601886e-06, |
| "loss": 2.8052, |
| "step": 884 |
| }, |
| { |
| "epoch": 0.885, |
| "grad_norm": 0.53125, |
| "learning_rate": 8.086466074476563e-06, |
| "loss": 2.7368, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.886, |
| "grad_norm": 0.55859375, |
| "learning_rate": 7.949514654755962e-06, |
| "loss": 2.6757, |
| "step": 886 |
| }, |
| { |
| "epoch": 0.887, |
| "grad_norm": 0.5546875, |
| "learning_rate": 7.81368484114996e-06, |
| "loss": 2.703, |
| "step": 887 |
| }, |
| { |
| "epoch": 0.888, |
| "grad_norm": 0.54296875, |
| "learning_rate": 7.67897828870191e-06, |
| "loss": 2.7158, |
| "step": 888 |
| }, |
| { |
| "epoch": 0.889, |
| "grad_norm": 0.55078125, |
| "learning_rate": 7.545396638768698e-06, |
| "loss": 2.7778, |
| "step": 889 |
| }, |
| { |
| "epoch": 0.89, |
| "grad_norm": 0.5078125, |
| "learning_rate": 7.412941519000527e-06, |
| "loss": 2.6752, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.891, |
| "grad_norm": 1.59375, |
| "learning_rate": 7.281614543321269e-06, |
| "loss": 2.7462, |
| "step": 891 |
| }, |
| { |
| "epoch": 0.892, |
| "grad_norm": 0.515625, |
| "learning_rate": 7.151417311908648e-06, |
| "loss": 2.6856, |
| "step": 892 |
| }, |
| { |
| "epoch": 0.893, |
| "grad_norm": 0.546875, |
| "learning_rate": 7.022351411174866e-06, |
| "loss": 2.7136, |
| "step": 893 |
| }, |
| { |
| "epoch": 0.894, |
| "grad_norm": 0.55859375, |
| "learning_rate": 6.894418413747183e-06, |
| "loss": 2.7652, |
| "step": 894 |
| }, |
| { |
| "epoch": 0.895, |
| "grad_norm": 0.5390625, |
| "learning_rate": 6.767619878448783e-06, |
| "loss": 2.687, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.896, |
| "grad_norm": 0.609375, |
| "learning_rate": 6.6419573502798374e-06, |
| "loss": 2.8021, |
| "step": 896 |
| }, |
| { |
| "epoch": 0.897, |
| "grad_norm": 0.53125, |
| "learning_rate": 6.517432360398556e-06, |
| "loss": 2.7689, |
| "step": 897 |
| }, |
| { |
| "epoch": 0.898, |
| "grad_norm": 0.5859375, |
| "learning_rate": 6.394046426102674e-06, |
| "loss": 2.7352, |
| "step": 898 |
| }, |
| { |
| "epoch": 0.899, |
| "grad_norm": 0.5546875, |
| "learning_rate": 6.2718010508108545e-06, |
| "loss": 2.7, |
| "step": 899 |
| }, |
| { |
| "epoch": 0.9, |
| "grad_norm": 0.55859375, |
| "learning_rate": 6.1506977240444074e-06, |
| "loss": 2.6995, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.901, |
| "grad_norm": 0.5390625, |
| "learning_rate": 6.030737921409169e-06, |
| "loss": 2.6935, |
| "step": 901 |
| }, |
| { |
| "epoch": 0.902, |
| "grad_norm": 0.5234375, |
| "learning_rate": 5.911923104577455e-06, |
| "loss": 2.711, |
| "step": 902 |
| }, |
| { |
| "epoch": 0.903, |
| "grad_norm": 0.5078125, |
| "learning_rate": 5.7942547212703315e-06, |
| "loss": 2.7778, |
| "step": 903 |
| }, |
| { |
| "epoch": 0.904, |
| "grad_norm": 0.53125, |
| "learning_rate": 5.6777342052399045e-06, |
| "loss": 2.7373, |
| "step": 904 |
| }, |
| { |
| "epoch": 0.905, |
| "grad_norm": 0.515625, |
| "learning_rate": 5.562362976251901e-06, |
| "loss": 2.7644, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.906, |
| "grad_norm": 0.54296875, |
| "learning_rate": 5.448142440068316e-06, |
| "loss": 2.6753, |
| "step": 906 |
| }, |
| { |
| "epoch": 0.907, |
| "grad_norm": 0.8515625, |
| "learning_rate": 5.335073988430372e-06, |
| "loss": 2.7881, |
| "step": 907 |
| }, |
| { |
| "epoch": 0.908, |
| "grad_norm": 0.5234375, |
| "learning_rate": 5.223158999041444e-06, |
| "loss": 2.8192, |
| "step": 908 |
| }, |
| { |
| "epoch": 0.909, |
| "grad_norm": 0.578125, |
| "learning_rate": 5.1123988355503475e-06, |
| "loss": 2.7979, |
| "step": 909 |
| }, |
| { |
| "epoch": 0.91, |
| "grad_norm": 0.578125, |
| "learning_rate": 5.002794847534764e-06, |
| "loss": 2.7629, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.911, |
| "grad_norm": 0.53515625, |
| "learning_rate": 4.8943483704846475e-06, |
| "loss": 2.7704, |
| "step": 911 |
| }, |
| { |
| "epoch": 0.912, |
| "grad_norm": 0.55859375, |
| "learning_rate": 4.7870607257861415e-06, |
| "loss": 2.6811, |
| "step": 912 |
| }, |
| { |
| "epoch": 0.913, |
| "grad_norm": 0.515625, |
| "learning_rate": 4.680933220705308e-06, |
| "loss": 2.7518, |
| "step": 913 |
| }, |
| { |
| "epoch": 0.914, |
| "grad_norm": 0.53125, |
| "learning_rate": 4.575967148372317e-06, |
| "loss": 2.7595, |
| "step": 914 |
| }, |
| { |
| "epoch": 0.915, |
| "grad_norm": 0.54296875, |
| "learning_rate": 4.4721637877656375e-06, |
| "loss": 2.7555, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.916, |
| "grad_norm": 0.546875, |
| "learning_rate": 4.369524403696457e-06, |
| "loss": 2.728, |
| "step": 916 |
| }, |
| { |
| "epoch": 0.917, |
| "grad_norm": 0.546875, |
| "learning_rate": 4.268050246793276e-06, |
| "loss": 2.7264, |
| "step": 917 |
| }, |
| { |
| "epoch": 0.918, |
| "grad_norm": 0.54296875, |
| "learning_rate": 4.167742553486675e-06, |
| "loss": 2.7391, |
| "step": 918 |
| }, |
| { |
| "epoch": 0.919, |
| "grad_norm": 0.53515625, |
| "learning_rate": 4.068602545994249e-06, |
| "loss": 2.7269, |
| "step": 919 |
| }, |
| { |
| "epoch": 0.92, |
| "grad_norm": 0.51953125, |
| "learning_rate": 3.970631432305694e-06, |
| "loss": 2.7273, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.921, |
| "grad_norm": 0.53515625, |
| "learning_rate": 3.873830406168111e-06, |
| "loss": 2.8126, |
| "step": 921 |
| }, |
| { |
| "epoch": 0.922, |
| "grad_norm": 2.171875, |
| "learning_rate": 3.7782006470714616e-06, |
| "loss": 2.7532, |
| "step": 922 |
| }, |
| { |
| "epoch": 0.923, |
| "grad_norm": 0.53125, |
| "learning_rate": 3.68374332023419e-06, |
| "loss": 2.7923, |
| "step": 923 |
| }, |
| { |
| "epoch": 0.924, |
| "grad_norm": 0.52734375, |
| "learning_rate": 3.5904595765890005e-06, |
| "loss": 2.7534, |
| "step": 924 |
| }, |
| { |
| "epoch": 0.925, |
| "grad_norm": 0.5390625, |
| "learning_rate": 3.4983505527688586e-06, |
| "loss": 2.7706, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.926, |
| "grad_norm": 0.5390625, |
| "learning_rate": 3.40741737109318e-06, |
| "loss": 2.6821, |
| "step": 926 |
| }, |
| { |
| "epoch": 0.927, |
| "grad_norm": 0.56640625, |
| "learning_rate": 3.3176611395540626e-06, |
| "loss": 2.6992, |
| "step": 927 |
| }, |
| { |
| "epoch": 0.928, |
| "grad_norm": 0.5546875, |
| "learning_rate": 3.2290829518028862e-06, |
| "loss": 2.7818, |
| "step": 928 |
| }, |
| { |
| "epoch": 0.929, |
| "grad_norm": 0.56640625, |
| "learning_rate": 3.1416838871368924e-06, |
| "loss": 2.7538, |
| "step": 929 |
| }, |
| { |
| "epoch": 0.93, |
| "grad_norm": 0.5546875, |
| "learning_rate": 3.0554650104861136e-06, |
| "loss": 2.7118, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.931, |
| "grad_norm": 0.55859375, |
| "learning_rate": 2.970427372400353e-06, |
| "loss": 2.7252, |
| "step": 931 |
| }, |
| { |
| "epoch": 0.932, |
| "grad_norm": 0.546875, |
| "learning_rate": 2.8865720090364034e-06, |
| "loss": 2.8308, |
| "step": 932 |
| }, |
| { |
| "epoch": 0.933, |
| "grad_norm": 0.53125, |
| "learning_rate": 2.8038999421453826e-06, |
| "loss": 2.6437, |
| "step": 933 |
| }, |
| { |
| "epoch": 0.934, |
| "grad_norm": 0.50390625, |
| "learning_rate": 2.7224121790603517e-06, |
| "loss": 2.764, |
| "step": 934 |
| }, |
| { |
| "epoch": 0.935, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.6421097126839712e-06, |
| "loss": 2.6876, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.936, |
| "grad_norm": 0.55078125, |
| "learning_rate": 2.5629935214764865e-06, |
| "loss": 2.8505, |
| "step": 936 |
| }, |
| { |
| "epoch": 0.937, |
| "grad_norm": 0.53515625, |
| "learning_rate": 2.4850645694436736e-06, |
| "loss": 2.7782, |
| "step": 937 |
| }, |
| { |
| "epoch": 0.938, |
| "grad_norm": 1.0625, |
| "learning_rate": 2.4083238061252567e-06, |
| "loss": 2.6892, |
| "step": 938 |
| }, |
| { |
| "epoch": 0.939, |
| "grad_norm": 0.52734375, |
| "learning_rate": 2.332772166583208e-06, |
| "loss": 2.6548, |
| "step": 939 |
| }, |
| { |
| "epoch": 0.94, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.2584105713904125e-06, |
| "loss": 2.7622, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.941, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.1852399266194314e-06, |
| "loss": 2.7183, |
| "step": 941 |
| }, |
| { |
| "epoch": 0.942, |
| "grad_norm": 0.5234375, |
| "learning_rate": 2.1132611238315003e-06, |
| "loss": 2.7956, |
| "step": 942 |
| }, |
| { |
| "epoch": 0.943, |
| "grad_norm": 0.50390625, |
| "learning_rate": 2.0424750400655947e-06, |
| "loss": 2.6357, |
| "step": 943 |
| }, |
| { |
| "epoch": 0.944, |
| "grad_norm": 0.55859375, |
| "learning_rate": 1.9728825378278246e-06, |
| "loss": 2.6503, |
| "step": 944 |
| }, |
| { |
| "epoch": 0.945, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.904484465080847e-06, |
| "loss": 2.6973, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.946, |
| "grad_norm": 0.546875, |
| "learning_rate": 1.8372816552336026e-06, |
| "loss": 2.7222, |
| "step": 946 |
| }, |
| { |
| "epoch": 0.947, |
| "grad_norm": 0.515625, |
| "learning_rate": 1.771274927131139e-06, |
| "loss": 2.6586, |
| "step": 947 |
| }, |
| { |
| "epoch": 0.948, |
| "grad_norm": 0.53125, |
| "learning_rate": 1.706465085044584e-06, |
| "loss": 2.657, |
| "step": 948 |
| }, |
| { |
| "epoch": 0.949, |
| "grad_norm": 0.55078125, |
| "learning_rate": 1.6428529186614195e-06, |
| "loss": 2.7474, |
| "step": 949 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.578125, |
| "learning_rate": 1.580439203075812e-06, |
| "loss": 2.6946, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.951, |
| "grad_norm": 0.5234375, |
| "learning_rate": 1.5192246987791981e-06, |
| "loss": 2.6957, |
| "step": 951 |
| }, |
| { |
| "epoch": 0.952, |
| "grad_norm": 0.53125, |
| "learning_rate": 1.4592101516509914e-06, |
| "loss": 2.7191, |
| "step": 952 |
| }, |
| { |
| "epoch": 0.953, |
| "grad_norm": 0.5234375, |
| "learning_rate": 1.400396292949513e-06, |
| "loss": 2.6774, |
| "step": 953 |
| }, |
| { |
| "epoch": 0.954, |
| "grad_norm": 0.640625, |
| "learning_rate": 1.3427838393030633e-06, |
| "loss": 2.785, |
| "step": 954 |
| }, |
| { |
| "epoch": 0.955, |
| "grad_norm": 0.578125, |
| "learning_rate": 1.2863734927012095e-06, |
| "loss": 2.7072, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.956, |
| "grad_norm": 0.55078125, |
| "learning_rate": 1.231165940486234e-06, |
| "loss": 2.7511, |
| "step": 956 |
| }, |
| { |
| "epoch": 0.957, |
| "grad_norm": 0.5546875, |
| "learning_rate": 1.1771618553447216e-06, |
| "loss": 2.7826, |
| "step": 957 |
| }, |
| { |
| "epoch": 0.958, |
| "grad_norm": 5.4375, |
| "learning_rate": 1.1243618952994195e-06, |
| "loss": 2.7722, |
| "step": 958 |
| }, |
| { |
| "epoch": 0.959, |
| "grad_norm": 0.50390625, |
| "learning_rate": 1.0727667037011668e-06, |
| "loss": 2.6715, |
| "step": 959 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 0.54296875, |
| "learning_rate": 1.0223769092211012e-06, |
| "loss": 2.7709, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.961, |
| "grad_norm": 0.5390625, |
| "learning_rate": 9.731931258429638e-07, |
| "loss": 2.7166, |
| "step": 961 |
| }, |
| { |
| "epoch": 0.962, |
| "grad_norm": 0.53515625, |
| "learning_rate": 9.252159528556403e-07, |
| "loss": 2.7073, |
| "step": 962 |
| }, |
| { |
| "epoch": 0.963, |
| "grad_norm": 0.5546875, |
| "learning_rate": 8.784459748458318e-07, |
| "loss": 2.6861, |
| "step": 963 |
| }, |
| { |
| "epoch": 0.964, |
| "grad_norm": 0.515625, |
| "learning_rate": 8.328837616909613e-07, |
| "loss": 2.7969, |
| "step": 964 |
| }, |
| { |
| "epoch": 0.965, |
| "grad_norm": 0.53125, |
| "learning_rate": 7.885298685522235e-07, |
| "loss": 2.6999, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.966, |
| "grad_norm": 0.56640625, |
| "learning_rate": 7.453848358678017e-07, |
| "loss": 2.7762, |
| "step": 966 |
| }, |
| { |
| "epoch": 0.967, |
| "grad_norm": 0.55078125, |
| "learning_rate": 7.034491893463058e-07, |
| "loss": 2.7231, |
| "step": 967 |
| }, |
| { |
| "epoch": 0.968, |
| "grad_norm": 0.515625, |
| "learning_rate": 6.627234399603555e-07, |
| "loss": 2.6857, |
| "step": 968 |
| }, |
| { |
| "epoch": 0.969, |
| "grad_norm": 2.28125, |
| "learning_rate": 6.232080839403631e-07, |
| "loss": 2.707, |
| "step": 969 |
| }, |
| { |
| "epoch": 0.97, |
| "grad_norm": 0.56640625, |
| "learning_rate": 5.849036027684606e-07, |
| "loss": 2.7359, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.971, |
| "grad_norm": 0.55078125, |
| "learning_rate": 5.478104631726711e-07, |
| "loss": 2.6725, |
| "step": 971 |
| }, |
| { |
| "epoch": 0.972, |
| "grad_norm": 0.53125, |
| "learning_rate": 5.119291171211793e-07, |
| "loss": 2.762, |
| "step": 972 |
| }, |
| { |
| "epoch": 0.973, |
| "grad_norm": 0.546875, |
| "learning_rate": 4.772600018168816e-07, |
| "loss": 2.7416, |
| "step": 973 |
| }, |
| { |
| "epoch": 0.974, |
| "grad_norm": 0.53125, |
| "learning_rate": 4.438035396920004e-07, |
| "loss": 2.7811, |
| "step": 974 |
| }, |
| { |
| "epoch": 0.975, |
| "grad_norm": 0.54296875, |
| "learning_rate": 4.115601384029666e-07, |
| "loss": 2.7719, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.976, |
| "grad_norm": 0.54296875, |
| "learning_rate": 3.805301908254455e-07, |
| "loss": 2.7054, |
| "step": 976 |
| }, |
| { |
| "epoch": 0.977, |
| "grad_norm": 0.546875, |
| "learning_rate": 3.50714075049563e-07, |
| "loss": 2.8006, |
| "step": 977 |
| }, |
| { |
| "epoch": 0.978, |
| "grad_norm": 0.58203125, |
| "learning_rate": 3.2211215437528694e-07, |
| "loss": 2.7788, |
| "step": 978 |
| }, |
| { |
| "epoch": 0.979, |
| "grad_norm": 0.53125, |
| "learning_rate": 2.947247773079753e-07, |
| "loss": 2.6917, |
| "step": 979 |
| }, |
| { |
| "epoch": 0.98, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.685522775541904e-07, |
| "loss": 2.7921, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.981, |
| "grad_norm": 0.58984375, |
| "learning_rate": 2.4359497401758024e-07, |
| "loss": 2.6928, |
| "step": 981 |
| }, |
| { |
| "epoch": 0.982, |
| "grad_norm": 0.515625, |
| "learning_rate": 2.1985317079500356e-07, |
| "loss": 2.8196, |
| "step": 982 |
| }, |
| { |
| "epoch": 0.983, |
| "grad_norm": 0.5546875, |
| "learning_rate": 1.973271571728441e-07, |
| "loss": 2.6882, |
| "step": 983 |
| }, |
| { |
| "epoch": 0.984, |
| "grad_norm": 0.546875, |
| "learning_rate": 1.7601720762346897e-07, |
| "loss": 2.7932, |
| "step": 984 |
| }, |
| { |
| "epoch": 0.985, |
| "grad_norm": 0.79296875, |
| "learning_rate": 1.5592358180189782e-07, |
| "loss": 2.7674, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.986, |
| "grad_norm": 0.5, |
| "learning_rate": 1.3704652454261668e-07, |
| "loss": 2.7227, |
| "step": 986 |
| }, |
| { |
| "epoch": 0.987, |
| "grad_norm": 0.5859375, |
| "learning_rate": 1.193862658566025e-07, |
| "loss": 2.6739, |
| "step": 987 |
| }, |
| { |
| "epoch": 0.988, |
| "grad_norm": 0.52734375, |
| "learning_rate": 1.0294302092853647e-07, |
| "loss": 2.731, |
| "step": 988 |
| }, |
| { |
| "epoch": 0.989, |
| "grad_norm": 0.578125, |
| "learning_rate": 8.771699011416168e-08, |
| "loss": 2.8035, |
| "step": 989 |
| }, |
| { |
| "epoch": 0.99, |
| "grad_norm": 0.53515625, |
| "learning_rate": 7.370835893788508e-08, |
| "loss": 2.7294, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.991, |
| "grad_norm": 0.55078125, |
| "learning_rate": 6.09172980904238e-08, |
| "loss": 2.8871, |
| "step": 991 |
| }, |
| { |
| "epoch": 0.992, |
| "grad_norm": 0.5390625, |
| "learning_rate": 4.934396342684e-08, |
| "loss": 2.7629, |
| "step": 992 |
| }, |
| { |
| "epoch": 0.993, |
| "grad_norm": 0.5859375, |
| "learning_rate": 3.898849596456478e-08, |
| "loss": 2.7404, |
| "step": 993 |
| }, |
| { |
| "epoch": 0.994, |
| "grad_norm": 0.5546875, |
| "learning_rate": 2.985102188168831e-08, |
| "loss": 2.7835, |
| "step": 994 |
| }, |
| { |
| "epoch": 0.995, |
| "grad_norm": 0.54296875, |
| "learning_rate": 2.193165251545004e-08, |
| "loss": 2.7149, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.996, |
| "grad_norm": 0.515625, |
| "learning_rate": 1.5230484360873044e-08, |
| "loss": 2.7078, |
| "step": 996 |
| }, |
| { |
| "epoch": 0.997, |
| "grad_norm": 0.53515625, |
| "learning_rate": 9.747599069576119e-09, |
| "loss": 2.6312, |
| "step": 997 |
| }, |
| { |
| "epoch": 0.998, |
| "grad_norm": 0.5078125, |
| "learning_rate": 5.483063448785686e-09, |
| "loss": 2.7117, |
| "step": 998 |
| }, |
| { |
| "epoch": 0.999, |
| "grad_norm": 0.53515625, |
| "learning_rate": 2.4369294605253166e-09, |
| "loss": 2.7865, |
| "step": 999 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.5859375, |
| "learning_rate": 6.092342209607083e-10, |
| "loss": 2.7628, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 2.704561948776245, |
| "eval_runtime": 31.803, |
| "eval_samples_per_second": 15.722, |
| "eval_steps_per_second": 2.641, |
| "step": 1000 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 1000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 9223372036854775807, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.818479221014528e+18, |
| "train_batch_size": 6, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|