| {"current_steps": 1, "total_steps": 168, "loss": 2.1358, "learning_rate": 9.999125804563732e-05, "epoch": 0.06779661016949153, "percentage": 0.6, "elapsed_time": "0:00:11", "remaining_time": "0:32:12", "throughput": 941.52, "total_tokens": 10896} |
| {"current_steps": 2, "total_steps": 168, "loss": 2.3575, "learning_rate": 9.996503523941994e-05, "epoch": 0.13559322033898305, "percentage": 1.19, "elapsed_time": "0:00:21", "remaining_time": "0:29:03", "throughput": 976.62, "total_tokens": 20512} |
| {"current_steps": 3, "total_steps": 168, "loss": 1.4712, "learning_rate": 9.992134075089084e-05, "epoch": 0.2033898305084746, "percentage": 1.79, "elapsed_time": "0:00:30", "remaining_time": "0:27:45", "throughput": 990.02, "total_tokens": 29984} |
| {"current_steps": 4, "total_steps": 168, "loss": 2.3124, "learning_rate": 9.986018985905901e-05, "epoch": 0.2711864406779661, "percentage": 2.38, "elapsed_time": "0:00:39", "remaining_time": "0:27:18", "throughput": 994.84, "total_tokens": 39760} |
| {"current_steps": 5, "total_steps": 168, "loss": 1.8062, "learning_rate": 9.978160394705668e-05, "epoch": 0.3389830508474576, "percentage": 2.98, "elapsed_time": "0:00:49", "remaining_time": "0:27:09", "throughput": 997.71, "total_tokens": 49856} |
| {"current_steps": 6, "total_steps": 168, "loss": 2.0499, "learning_rate": 9.968561049466214e-05, "epoch": 0.4067796610169492, "percentage": 3.57, "elapsed_time": "0:01:00", "remaining_time": "0:27:07", "throughput": 999.12, "total_tokens": 60208} |
| {"current_steps": 7, "total_steps": 168, "loss": 1.1603, "learning_rate": 9.957224306869053e-05, "epoch": 0.4745762711864407, "percentage": 4.17, "elapsed_time": "0:01:10", "remaining_time": "0:27:12", "throughput": 997.92, "total_tokens": 70816} |
| {"current_steps": 8, "total_steps": 168, "loss": 1.9923, "learning_rate": 9.944154131125642e-05, "epoch": 0.5423728813559322, "percentage": 4.76, "elapsed_time": "0:01:19", "remaining_time": "0:26:37", "throughput": 1000.97, "total_tokens": 79968} |
| {"current_steps": 9, "total_steps": 168, "loss": 1.9688, "learning_rate": 9.92935509259118e-05, "epoch": 0.6101694915254238, "percentage": 5.36, "elapsed_time": "0:01:29", "remaining_time": "0:26:17", "throughput": 1002.08, "total_tokens": 89472} |
| {"current_steps": 10, "total_steps": 168, "loss": 2.2878, "learning_rate": 9.912832366166442e-05, "epoch": 0.6779661016949152, "percentage": 5.95, "elapsed_time": "0:01:40", "remaining_time": "0:26:25", "throughput": 999.99, "total_tokens": 100336} |
| {"current_steps": 11, "total_steps": 168, "loss": 2.001, "learning_rate": 9.894591729488242e-05, "epoch": 0.7457627118644068, "percentage": 6.55, "elapsed_time": "0:01:51", "remaining_time": "0:26:30", "throughput": 992.58, "total_tokens": 110640} |
| {"current_steps": 12, "total_steps": 168, "loss": 1.8458, "learning_rate": 9.874639560909117e-05, "epoch": 0.8135593220338984, "percentage": 7.14, "elapsed_time": "0:02:01", "remaining_time": "0:26:13", "throughput": 993.78, "total_tokens": 120304} |
| {"current_steps": 13, "total_steps": 168, "loss": 1.8775, "learning_rate": 9.852982837266955e-05, "epoch": 0.8813559322033898, "percentage": 7.74, "elapsed_time": "0:02:08", "remaining_time": "0:25:35", "throughput": 996.11, "total_tokens": 128288} |
| {"current_steps": 14, "total_steps": 168, "loss": 1.9309, "learning_rate": 9.829629131445342e-05, "epoch": 0.9491525423728814, "percentage": 8.33, "elapsed_time": "0:02:19", "remaining_time": "0:25:30", "throughput": 996.64, "total_tokens": 138624} |
| {"current_steps": 15, "total_steps": 168, "loss": 1.5507, "learning_rate": 9.804586609725499e-05, "epoch": 1.0169491525423728, "percentage": 8.93, "elapsed_time": "0:02:28", "remaining_time": "0:25:18", "throughput": 995.58, "total_tokens": 148208} |
| {"current_steps": 16, "total_steps": 168, "loss": 1.6802, "learning_rate": 9.777864028930705e-05, "epoch": 1.0847457627118644, "percentage": 9.52, "elapsed_time": "0:02:39", "remaining_time": "0:25:18", "throughput": 995.84, "total_tokens": 159200} |
| {"current_steps": 17, "total_steps": 168, "loss": 1.5648, "learning_rate": 9.74947073336423e-05, "epoch": 1.152542372881356, "percentage": 10.12, "elapsed_time": "0:02:46", "remaining_time": "0:24:42", "throughput": 997.57, "total_tokens": 166544} |
| {"current_steps": 18, "total_steps": 168, "loss": 1.3598, "learning_rate": 9.719416651541839e-05, "epoch": 1.2203389830508475, "percentage": 10.71, "elapsed_time": "0:02:57", "remaining_time": "0:24:40", "throughput": 996.63, "total_tokens": 177120} |
| {"current_steps": 19, "total_steps": 168, "loss": 1.3238, "learning_rate": 9.687712292719997e-05, "epoch": 1.288135593220339, "percentage": 11.31, "elapsed_time": "0:03:06", "remaining_time": "0:24:23", "throughput": 998.15, "total_tokens": 186272} |
| {"current_steps": 20, "total_steps": 168, "loss": 1.8704, "learning_rate": 9.654368743221022e-05, "epoch": 1.3559322033898304, "percentage": 11.9, "elapsed_time": "0:03:16", "remaining_time": "0:24:15", "throughput": 999.09, "total_tokens": 196480} |
| {"current_steps": 21, "total_steps": 168, "loss": 1.8205, "learning_rate": 9.619397662556435e-05, "epoch": 1.423728813559322, "percentage": 12.5, "elapsed_time": "0:03:24", "remaining_time": "0:23:50", "throughput": 997.64, "total_tokens": 203856} |
| {"current_steps": 22, "total_steps": 168, "loss": 1.3595, "learning_rate": 9.582811279349882e-05, "epoch": 1.4915254237288136, "percentage": 13.1, "elapsed_time": "0:03:35", "remaining_time": "0:23:47", "throughput": 997.51, "total_tokens": 214512} |
| {"current_steps": 23, "total_steps": 168, "loss": 2.0797, "learning_rate": 9.544622387061055e-05, "epoch": 1.559322033898305, "percentage": 13.69, "elapsed_time": "0:03:44", "remaining_time": "0:23:33", "throughput": 998.42, "total_tokens": 223792} |
| {"current_steps": 24, "total_steps": 168, "loss": 2.15, "learning_rate": 9.504844339512095e-05, "epoch": 1.6271186440677967, "percentage": 14.29, "elapsed_time": "0:03:54", "remaining_time": "0:23:24", "throughput": 998.86, "total_tokens": 233776} |
| {"current_steps": 25, "total_steps": 168, "loss": 2.0059, "learning_rate": 9.463491046218058e-05, "epoch": 1.694915254237288, "percentage": 14.88, "elapsed_time": "0:04:03", "remaining_time": "0:23:13", "throughput": 999.83, "total_tokens": 243600} |
| {"current_steps": 26, "total_steps": 168, "loss": 1.2697, "learning_rate": 9.420576967523049e-05, "epoch": 1.7627118644067796, "percentage": 15.48, "elapsed_time": "0:04:14", "remaining_time": "0:23:11", "throughput": 999.23, "total_tokens": 254608} |
| {"current_steps": 27, "total_steps": 168, "loss": 1.6259, "learning_rate": 9.376117109543769e-05, "epoch": 1.8305084745762712, "percentage": 16.07, "elapsed_time": "0:04:26", "remaining_time": "0:23:09", "throughput": 998.72, "total_tokens": 265728} |
| {"current_steps": 28, "total_steps": 168, "loss": 1.0618, "learning_rate": 9.330127018922194e-05, "epoch": 1.8983050847457628, "percentage": 16.67, "elapsed_time": "0:04:37", "remaining_time": "0:23:05", "throughput": 998.28, "total_tokens": 276624} |
| {"current_steps": 29, "total_steps": 168, "loss": 1.8557, "learning_rate": 9.282622777389258e-05, "epoch": 1.9661016949152543, "percentage": 17.26, "elapsed_time": "0:04:47", "remaining_time": "0:22:58", "throughput": 998.39, "total_tokens": 287072} |
| {"current_steps": 30, "total_steps": 168, "loss": 1.8077, "learning_rate": 9.233620996141421e-05, "epoch": 2.0338983050847457, "percentage": 17.86, "elapsed_time": "0:04:56", "remaining_time": "0:22:45", "throughput": 998.09, "total_tokens": 296248} |
| {"current_steps": 31, "total_steps": 168, "loss": 1.767, "learning_rate": 9.183138810032099e-05, "epoch": 2.1016949152542375, "percentage": 18.45, "elapsed_time": "0:05:06", "remaining_time": "0:22:33", "throughput": 997.48, "total_tokens": 305560} |
| {"current_steps": 32, "total_steps": 168, "loss": 1.5561, "learning_rate": 9.131193871579975e-05, "epoch": 2.169491525423729, "percentage": 19.05, "elapsed_time": "0:05:18", "remaining_time": "0:22:32", "throughput": 996.38, "total_tokens": 317176} |
| {"current_steps": 33, "total_steps": 168, "loss": 1.7631, "learning_rate": 9.077804344796302e-05, "epoch": 2.23728813559322, "percentage": 19.64, "elapsed_time": "0:05:28", "remaining_time": "0:22:24", "throughput": 995.95, "total_tokens": 327352} |
| {"current_steps": 34, "total_steps": 168, "loss": 1.5754, "learning_rate": 9.022988898833342e-05, "epoch": 2.305084745762712, "percentage": 20.24, "elapsed_time": "0:05:38", "remaining_time": "0:22:15", "throughput": 995.95, "total_tokens": 337384} |
| {"current_steps": 35, "total_steps": 168, "loss": 1.4501, "learning_rate": 8.966766701456177e-05, "epoch": 2.3728813559322033, "percentage": 20.83, "elapsed_time": "0:05:48", "remaining_time": "0:22:04", "throughput": 995.46, "total_tokens": 346952} |
| {"current_steps": 36, "total_steps": 168, "loss": 1.8493, "learning_rate": 8.90915741234015e-05, "epoch": 2.440677966101695, "percentage": 21.43, "elapsed_time": "0:05:58", "remaining_time": "0:21:53", "throughput": 995.34, "total_tokens": 356568} |
| {"current_steps": 37, "total_steps": 168, "loss": 1.6969, "learning_rate": 8.850181176196315e-05, "epoch": 2.5084745762711864, "percentage": 22.02, "elapsed_time": "0:06:08", "remaining_time": "0:21:43", "throughput": 995.19, "total_tokens": 366504} |
| {"current_steps": 38, "total_steps": 168, "loss": 1.3998, "learning_rate": 8.789858615727265e-05, "epoch": 2.576271186440678, "percentage": 22.62, "elapsed_time": "0:06:19", "remaining_time": "0:21:38", "throughput": 994.4, "total_tokens": 377512} |
| {"current_steps": 39, "total_steps": 168, "loss": 1.1354, "learning_rate": 8.728210824415827e-05, "epoch": 2.6440677966101696, "percentage": 23.21, "elapsed_time": "0:06:30", "remaining_time": "0:21:33", "throughput": 994.35, "total_tokens": 388728} |
| {"current_steps": 40, "total_steps": 168, "loss": 1.5526, "learning_rate": 8.665259359149132e-05, "epoch": 2.711864406779661, "percentage": 23.81, "elapsed_time": "0:06:35", "remaining_time": "0:21:06", "throughput": 994.95, "total_tokens": 393800} |
| {"current_steps": 41, "total_steps": 168, "loss": 1.671, "learning_rate": 8.601026232680634e-05, "epoch": 2.7796610169491527, "percentage": 24.4, "elapsed_time": "0:06:46", "remaining_time": "0:20:59", "throughput": 993.65, "total_tokens": 404072} |
| {"current_steps": 42, "total_steps": 168, "loss": 1.1851, "learning_rate": 8.535533905932738e-05, "epoch": 2.847457627118644, "percentage": 25.0, "elapsed_time": "0:06:58", "remaining_time": "0:20:54", "throughput": 993.5, "total_tokens": 415336} |
| {"current_steps": 43, "total_steps": 168, "loss": 1.8113, "learning_rate": 8.468805280142709e-05, "epoch": 2.915254237288136, "percentage": 25.6, "elapsed_time": "0:07:07", "remaining_time": "0:20:44", "throughput": 993.72, "total_tokens": 425288} |
| {"current_steps": 44, "total_steps": 168, "loss": 1.2456, "learning_rate": 8.400863688854597e-05, "epoch": 2.983050847457627, "percentage": 26.19, "elapsed_time": "0:07:19", "remaining_time": "0:20:38", "throughput": 993.61, "total_tokens": 436584} |
| {"current_steps": 45, "total_steps": 168, "loss": 1.479, "learning_rate": 8.33173288976002e-05, "epoch": 3.0508474576271185, "percentage": 26.79, "elapsed_time": "0:07:27", "remaining_time": "0:20:24", "throughput": 993.64, "total_tokens": 445072} |
| {"current_steps": 46, "total_steps": 168, "loss": 1.8757, "learning_rate": 8.261437056390606e-05, "epoch": 3.1186440677966103, "percentage": 27.38, "elapsed_time": "0:07:36", "remaining_time": "0:20:10", "throughput": 994.44, "total_tokens": 453696} |
| {"current_steps": 47, "total_steps": 168, "loss": 1.4613, "learning_rate": 8.190000769665044e-05, "epoch": 3.1864406779661016, "percentage": 27.98, "elapsed_time": "0:07:45", "remaining_time": "0:19:58", "throughput": 994.74, "total_tokens": 462976} |
| {"current_steps": 48, "total_steps": 168, "loss": 1.2657, "learning_rate": 8.117449009293668e-05, "epoch": 3.2542372881355934, "percentage": 28.57, "elapsed_time": "0:07:56", "remaining_time": "0:19:51", "throughput": 994.54, "total_tokens": 474016} |
| {"current_steps": 49, "total_steps": 168, "loss": 1.7769, "learning_rate": 8.043807145043604e-05, "epoch": 3.3220338983050848, "percentage": 29.17, "elapsed_time": "0:08:05", "remaining_time": "0:19:39", "throughput": 994.69, "total_tokens": 482960} |
| {"current_steps": 50, "total_steps": 168, "loss": 1.3444, "learning_rate": 7.969100927867507e-05, "epoch": 3.389830508474576, "percentage": 29.76, "elapsed_time": "0:08:15", "remaining_time": "0:19:29", "throughput": 995.02, "total_tokens": 493200} |
| {"current_steps": 51, "total_steps": 168, "loss": 1.4113, "learning_rate": 7.89335648089903e-05, "epoch": 3.457627118644068, "percentage": 30.36, "elapsed_time": "0:08:25", "remaining_time": "0:19:20", "throughput": 994.41, "total_tokens": 503168} |
| {"current_steps": 52, "total_steps": 168, "loss": 1.5596, "learning_rate": 7.81660029031811e-05, "epoch": 3.5254237288135593, "percentage": 30.95, "elapsed_time": "0:08:36", "remaining_time": "0:19:12", "throughput": 994.37, "total_tokens": 513776} |
| {"current_steps": 53, "total_steps": 168, "loss": 1.7417, "learning_rate": 7.738859196089358e-05, "epoch": 3.593220338983051, "percentage": 31.55, "elapsed_time": "0:08:43", "remaining_time": "0:18:56", "throughput": 995.02, "total_tokens": 521376} |
| {"current_steps": 54, "total_steps": 168, "loss": 1.7439, "learning_rate": 7.660160382576683e-05, "epoch": 3.6610169491525424, "percentage": 32.14, "elapsed_time": "0:08:54", "remaining_time": "0:18:47", "throughput": 995.29, "total_tokens": 531504} |
| {"current_steps": 55, "total_steps": 168, "loss": 1.2392, "learning_rate": 7.580531369037533e-05, "epoch": 3.7288135593220337, "percentage": 32.74, "elapsed_time": "0:09:05", "remaining_time": "0:18:40", "throughput": 995.24, "total_tokens": 542832} |
| {"current_steps": 56, "total_steps": 168, "loss": 1.4463, "learning_rate": 7.500000000000001e-05, "epoch": 3.7966101694915255, "percentage": 33.33, "elapsed_time": "0:09:16", "remaining_time": "0:18:33", "throughput": 995.21, "total_tokens": 554128} |
| {"current_steps": 57, "total_steps": 168, "loss": 1.1846, "learning_rate": 7.4185944355262e-05, "epoch": 3.864406779661017, "percentage": 33.93, "elapsed_time": "0:09:27", "remaining_time": "0:18:24", "throughput": 995.41, "total_tokens": 564576} |
| {"current_steps": 58, "total_steps": 168, "loss": 1.1875, "learning_rate": 7.33634314136531e-05, "epoch": 3.9322033898305087, "percentage": 34.52, "elapsed_time": "0:09:36", "remaining_time": "0:18:14", "throughput": 995.68, "total_tokens": 574464} |
| {"current_steps": 59, "total_steps": 168, "loss": 1.1346, "learning_rate": 7.253274878999727e-05, "epoch": 4.0, "percentage": 35.12, "elapsed_time": "0:09:46", "remaining_time": "0:18:03", "throughput": 995.88, "total_tokens": 584192} |
| {"current_steps": 60, "total_steps": 168, "loss": 1.6462, "learning_rate": 7.169418695587791e-05, "epoch": 4.067796610169491, "percentage": 35.71, "elapsed_time": "0:09:55", "remaining_time": "0:17:52", "throughput": 996.36, "total_tokens": 593440} |
| {"current_steps": 61, "total_steps": 168, "loss": 1.0892, "learning_rate": 7.084803913806641e-05, "epoch": 4.135593220338983, "percentage": 36.31, "elapsed_time": "0:10:06", "remaining_time": "0:17:43", "throughput": 995.0, "total_tokens": 603168} |
| {"current_steps": 62, "total_steps": 168, "loss": 1.4054, "learning_rate": 6.999460121598704e-05, "epoch": 4.203389830508475, "percentage": 36.9, "elapsed_time": "0:10:14", "remaining_time": "0:17:31", "throughput": 995.36, "total_tokens": 611888} |
| {"current_steps": 63, "total_steps": 168, "loss": 0.9302, "learning_rate": 6.91341716182545e-05, "epoch": 4.271186440677966, "percentage": 37.5, "elapsed_time": "0:10:26", "remaining_time": "0:17:23", "throughput": 995.19, "total_tokens": 623312} |
| {"current_steps": 64, "total_steps": 168, "loss": 1.2218, "learning_rate": 6.826705121831976e-05, "epoch": 4.338983050847458, "percentage": 38.1, "elapsed_time": "0:10:37", "remaining_time": "0:17:16", "throughput": 995.15, "total_tokens": 634768} |
| {"current_steps": 65, "total_steps": 168, "loss": 1.1605, "learning_rate": 6.739354322926136e-05, "epoch": 4.406779661016949, "percentage": 38.69, "elapsed_time": "0:10:48", "remaining_time": "0:17:07", "throughput": 994.93, "total_tokens": 644848} |
| {"current_steps": 66, "total_steps": 168, "loss": 1.7099, "learning_rate": 6.651395309775837e-05, "epoch": 4.47457627118644, "percentage": 39.29, "elapsed_time": "0:10:57", "remaining_time": "0:16:56", "throughput": 995.24, "total_tokens": 654672} |
| {"current_steps": 67, "total_steps": 168, "loss": 0.8147, "learning_rate": 6.562858839728223e-05, "epoch": 4.5423728813559325, "percentage": 39.88, "elapsed_time": "0:11:08", "remaining_time": "0:16:47", "throughput": 995.44, "total_tokens": 665584} |
| {"current_steps": 68, "total_steps": 168, "loss": 1.5787, "learning_rate": 6.473775872054521e-05, "epoch": 4.610169491525424, "percentage": 40.48, "elapsed_time": "0:11:16", "remaining_time": "0:16:35", "throughput": 995.9, "total_tokens": 674096} |
| {"current_steps": 69, "total_steps": 168, "loss": 1.2265, "learning_rate": 6.384177557124247e-05, "epoch": 4.677966101694915, "percentage": 41.07, "elapsed_time": "0:11:26", "remaining_time": "0:16:25", "throughput": 995.98, "total_tokens": 683968} |
| {"current_steps": 70, "total_steps": 168, "loss": 1.2753, "learning_rate": 6.294095225512603e-05, "epoch": 4.745762711864407, "percentage": 41.67, "elapsed_time": "0:11:37", "remaining_time": "0:16:15", "throughput": 996.17, "total_tokens": 694368} |
| {"current_steps": 71, "total_steps": 168, "loss": 1.6676, "learning_rate": 6.203560377044866e-05, "epoch": 4.813559322033898, "percentage": 42.26, "elapsed_time": "0:11:46", "remaining_time": "0:16:05", "throughput": 995.79, "total_tokens": 703440} |
| {"current_steps": 72, "total_steps": 168, "loss": 1.4984, "learning_rate": 6.112604669781572e-05, "epoch": 4.88135593220339, "percentage": 42.86, "elapsed_time": "0:11:54", "remaining_time": "0:15:52", "throughput": 996.1, "total_tokens": 711264} |
| {"current_steps": 73, "total_steps": 168, "loss": 1.6282, "learning_rate": 6.021259908948402e-05, "epoch": 4.9491525423728815, "percentage": 43.45, "elapsed_time": "0:12:05", "remaining_time": "0:15:44", "throughput": 995.88, "total_tokens": 722624} |
| {"current_steps": 74, "total_steps": 168, "loss": 0.8916, "learning_rate": 5.9295580358145744e-05, "epoch": 5.016949152542373, "percentage": 44.05, "elapsed_time": "0:12:13", "remaining_time": "0:15:32", "throughput": 996.05, "total_tokens": 730976} |
| {"current_steps": 75, "total_steps": 168, "loss": 0.7409, "learning_rate": 5.837531116523682e-05, "epoch": 5.084745762711864, "percentage": 44.64, "elapsed_time": "0:12:22", "remaining_time": "0:15:21", "throughput": 996.36, "total_tokens": 740160} |
| {"current_steps": 76, "total_steps": 168, "loss": 1.3499, "learning_rate": 5.745211330880872e-05, "epoch": 5.1525423728813555, "percentage": 45.24, "elapsed_time": "0:12:33", "remaining_time": "0:15:11", "throughput": 996.32, "total_tokens": 750544} |
| {"current_steps": 77, "total_steps": 168, "loss": 0.901, "learning_rate": 5.6526309611002594e-05, "epoch": 5.220338983050848, "percentage": 45.83, "elapsed_time": "0:12:44", "remaining_time": "0:15:03", "throughput": 996.22, "total_tokens": 761744} |
| {"current_steps": 78, "total_steps": 168, "loss": 1.1879, "learning_rate": 5.559822380516539e-05, "epoch": 5.288135593220339, "percentage": 46.43, "elapsed_time": "0:12:55", "remaining_time": "0:14:55", "throughput": 996.17, "total_tokens": 772944} |
| {"current_steps": 79, "total_steps": 168, "loss": 1.0959, "learning_rate": 5.466818042264753e-05, "epoch": 5.3559322033898304, "percentage": 47.02, "elapsed_time": "0:13:03", "remaining_time": "0:14:43", "throughput": 996.61, "total_tokens": 781296} |
| {"current_steps": 80, "total_steps": 168, "loss": 1.2333, "learning_rate": 5.373650467932122e-05, "epoch": 5.423728813559322, "percentage": 47.62, "elapsed_time": "0:13:15", "remaining_time": "0:14:34", "throughput": 996.46, "total_tokens": 792416} |
| {"current_steps": 81, "total_steps": 168, "loss": 1.1665, "learning_rate": 5.2803522361859594e-05, "epoch": 5.491525423728813, "percentage": 48.21, "elapsed_time": "0:13:26", "remaining_time": "0:14:25", "throughput": 995.65, "total_tokens": 802688} |
| {"current_steps": 82, "total_steps": 168, "loss": 1.0339, "learning_rate": 5.18695597138163e-05, "epoch": 5.559322033898305, "percentage": 48.81, "elapsed_time": "0:13:36", "remaining_time": "0:14:16", "throughput": 995.77, "total_tokens": 813376} |
| {"current_steps": 83, "total_steps": 168, "loss": 1.0024, "learning_rate": 5.0934943321545115e-05, "epoch": 5.627118644067797, "percentage": 49.4, "elapsed_time": "0:13:46", "remaining_time": "0:14:05", "throughput": 995.96, "total_tokens": 822752} |
| {"current_steps": 84, "total_steps": 168, "loss": 0.9337, "learning_rate": 5e-05, "epoch": 5.694915254237288, "percentage": 50.0, "elapsed_time": "0:13:55", "remaining_time": "0:13:55", "throughput": 996.18, "total_tokens": 832720} |
| {"current_steps": 85, "total_steps": 168, "loss": 1.0879, "learning_rate": 4.9065056678454904e-05, "epoch": 5.762711864406779, "percentage": 50.6, "elapsed_time": "0:14:06", "remaining_time": "0:13:46", "throughput": 996.3, "total_tokens": 842880} |
| {"current_steps": 86, "total_steps": 168, "loss": 1.3822, "learning_rate": 4.813044028618373e-05, "epoch": 5.830508474576272, "percentage": 51.19, "elapsed_time": "0:14:14", "remaining_time": "0:13:35", "throughput": 996.53, "total_tokens": 851968} |
| {"current_steps": 87, "total_steps": 168, "loss": 1.3801, "learning_rate": 4.7196477638140404e-05, "epoch": 5.898305084745763, "percentage": 51.79, "elapsed_time": "0:14:24", "remaining_time": "0:13:24", "throughput": 996.83, "total_tokens": 861632} |
| {"current_steps": 88, "total_steps": 168, "loss": 1.3577, "learning_rate": 4.626349532067879e-05, "epoch": 5.966101694915254, "percentage": 52.38, "elapsed_time": "0:14:35", "remaining_time": "0:13:15", "throughput": 996.76, "total_tokens": 872352} |
| {"current_steps": 89, "total_steps": 168, "loss": 0.7063, "learning_rate": 4.5331819577352474e-05, "epoch": 6.033898305084746, "percentage": 52.98, "elapsed_time": "0:14:43", "remaining_time": "0:13:04", "throughput": 996.98, "total_tokens": 881072} |
| {"current_steps": 90, "total_steps": 168, "loss": 1.35, "learning_rate": 4.4401776194834613e-05, "epoch": 6.101694915254237, "percentage": 53.57, "elapsed_time": "0:14:54", "remaining_time": "0:12:55", "throughput": 996.93, "total_tokens": 891840} |
| {"current_steps": 91, "total_steps": 168, "loss": 1.2183, "learning_rate": 4.347369038899744e-05, "epoch": 6.169491525423728, "percentage": 54.17, "elapsed_time": "0:15:04", "remaining_time": "0:12:45", "throughput": 996.51, "total_tokens": 901232} |
| {"current_steps": 92, "total_steps": 168, "loss": 1.0497, "learning_rate": 4.254788669119127e-05, "epoch": 6.237288135593221, "percentage": 54.76, "elapsed_time": "0:15:13", "remaining_time": "0:12:34", "throughput": 996.59, "total_tokens": 910496} |
| {"current_steps": 93, "total_steps": 168, "loss": 0.6885, "learning_rate": 4.162468883476319e-05, "epoch": 6.305084745762712, "percentage": 55.36, "elapsed_time": "0:15:22", "remaining_time": "0:12:24", "throughput": 996.82, "total_tokens": 919856} |
| {"current_steps": 94, "total_steps": 168, "loss": 0.9335, "learning_rate": 4.0704419641854274e-05, "epoch": 6.372881355932203, "percentage": 55.95, "elapsed_time": "0:15:34", "remaining_time": "0:12:15", "throughput": 996.66, "total_tokens": 931104} |
| {"current_steps": 95, "total_steps": 168, "loss": 1.0183, "learning_rate": 3.978740091051599e-05, "epoch": 6.440677966101695, "percentage": 56.55, "elapsed_time": "0:15:43", "remaining_time": "0:12:05", "throughput": 996.89, "total_tokens": 940976} |
| {"current_steps": 96, "total_steps": 168, "loss": 1.0591, "learning_rate": 3.887395330218429e-05, "epoch": 6.508474576271187, "percentage": 57.14, "elapsed_time": "0:15:51", "remaining_time": "0:11:53", "throughput": 997.25, "total_tokens": 949248} |
| {"current_steps": 97, "total_steps": 168, "loss": 1.0101, "learning_rate": 3.7964396229551364e-05, "epoch": 6.576271186440678, "percentage": 57.74, "elapsed_time": "0:16:03", "remaining_time": "0:11:44", "throughput": 997.22, "total_tokens": 960336} |
| {"current_steps": 98, "total_steps": 168, "loss": 0.6923, "learning_rate": 3.705904774487396e-05, "epoch": 6.6440677966101696, "percentage": 58.33, "elapsed_time": "0:16:13", "remaining_time": "0:11:35", "throughput": 997.24, "total_tokens": 970480} |
| {"current_steps": 99, "total_steps": 168, "loss": 0.9599, "learning_rate": 3.6158224428757535e-05, "epoch": 6.711864406779661, "percentage": 58.93, "elapsed_time": "0:16:21", "remaining_time": "0:11:24", "throughput": 997.32, "total_tokens": 979104} |
| {"current_steps": 100, "total_steps": 168, "loss": 1.0911, "learning_rate": 3.5262241279454785e-05, "epoch": 6.779661016949152, "percentage": 59.52, "elapsed_time": "0:16:31", "remaining_time": "0:11:14", "throughput": 997.43, "total_tokens": 989312} |
| {"current_steps": 101, "total_steps": 168, "loss": 0.9246, "learning_rate": 3.4371411602717784e-05, "epoch": 6.847457627118644, "percentage": 60.12, "elapsed_time": "0:16:42", "remaining_time": "0:11:04", "throughput": 996.93, "total_tokens": 999312} |
| {"current_steps": 102, "total_steps": 168, "loss": 0.811, "learning_rate": 3.3486046902241664e-05, "epoch": 6.915254237288136, "percentage": 60.71, "elapsed_time": "0:16:53", "remaining_time": "0:10:55", "throughput": 996.87, "total_tokens": 1010096} |
| {"current_steps": 103, "total_steps": 168, "loss": 0.8122, "learning_rate": 3.2606456770738636e-05, "epoch": 6.983050847457627, "percentage": 61.31, "elapsed_time": "0:17:01", "remaining_time": "0:10:44", "throughput": 996.91, "total_tokens": 1018704} |
| {"current_steps": 104, "total_steps": 168, "loss": 0.6868, "learning_rate": 3.173294878168025e-05, "epoch": 7.0508474576271185, "percentage": 61.9, "elapsed_time": "0:17:10", "remaining_time": "0:10:34", "throughput": 996.81, "total_tokens": 1027008} |
| {"current_steps": 105, "total_steps": 168, "loss": 0.8705, "learning_rate": 3.086582838174551e-05, "epoch": 7.11864406779661, "percentage": 62.5, "elapsed_time": "0:17:21", "remaining_time": "0:10:24", "throughput": 996.8, "total_tokens": 1038016} |
| {"current_steps": 106, "total_steps": 168, "loss": 1.0279, "learning_rate": 3.000539878401296e-05, "epoch": 7.186440677966102, "percentage": 63.1, "elapsed_time": "0:17:30", "remaining_time": "0:10:14", "throughput": 997.01, "total_tokens": 1046912} |
| {"current_steps": 107, "total_steps": 168, "loss": 0.9108, "learning_rate": 2.9151960861933614e-05, "epoch": 7.254237288135593, "percentage": 63.69, "elapsed_time": "0:17:39", "remaining_time": "0:10:03", "throughput": 997.19, "total_tokens": 1056480} |
| {"current_steps": 108, "total_steps": 168, "loss": 0.9454, "learning_rate": 2.8305813044122097e-05, "epoch": 7.322033898305085, "percentage": 64.29, "elapsed_time": "0:17:48", "remaining_time": "0:09:53", "throughput": 997.48, "total_tokens": 1066080} |
| {"current_steps": 109, "total_steps": 168, "loss": 0.7537, "learning_rate": 2.746725121000273e-05, "epoch": 7.389830508474576, "percentage": 64.88, "elapsed_time": "0:17:57", "remaining_time": "0:09:43", "throughput": 997.72, "total_tokens": 1075376} |
| {"current_steps": 110, "total_steps": 168, "loss": 0.87, "learning_rate": 2.66365685863469e-05, "epoch": 7.4576271186440675, "percentage": 65.48, "elapsed_time": "0:18:08", "remaining_time": "0:09:34", "throughput": 997.68, "total_tokens": 1086448} |
| {"current_steps": 111, "total_steps": 168, "loss": 0.5695, "learning_rate": 2.581405564473801e-05, "epoch": 7.52542372881356, "percentage": 66.07, "elapsed_time": "0:18:21", "remaining_time": "0:09:25", "throughput": 996.84, "total_tokens": 1098192} |
| {"current_steps": 112, "total_steps": 168, "loss": 0.5695, "learning_rate": 2.500000000000001e-05, "epoch": 7.593220338983051, "percentage": 66.67, "elapsed_time": "0:18:32", "remaining_time": "0:09:16", "throughput": 996.81, "total_tokens": 1108912} |
| {"current_steps": 113, "total_steps": 168, "loss": 0.7164, "learning_rate": 2.4194686309624663e-05, "epoch": 7.661016949152542, "percentage": 67.26, "elapsed_time": "0:18:40", "remaining_time": "0:09:05", "throughput": 996.92, "total_tokens": 1116672} |
| {"current_steps": 114, "total_steps": 168, "loss": 0.8342, "learning_rate": 2.3398396174233178e-05, "epoch": 7.728813559322034, "percentage": 67.86, "elapsed_time": "0:18:50", "remaining_time": "0:08:55", "throughput": 996.96, "total_tokens": 1126640} |
| {"current_steps": 115, "total_steps": 168, "loss": 0.8934, "learning_rate": 2.261140803910644e-05, "epoch": 7.796610169491525, "percentage": 68.45, "elapsed_time": "0:18:59", "remaining_time": "0:08:45", "throughput": 997.13, "total_tokens": 1135920} |
| {"current_steps": 116, "total_steps": 168, "loss": 0.8654, "learning_rate": 2.1833997096818898e-05, "epoch": 7.864406779661017, "percentage": 69.05, "elapsed_time": "0:19:10", "remaining_time": "0:08:35", "throughput": 997.08, "total_tokens": 1147216} |
| {"current_steps": 117, "total_steps": 168, "loss": 0.4848, "learning_rate": 2.1066435191009715e-05, "epoch": 7.932203389830509, "percentage": 69.64, "elapsed_time": "0:19:21", "remaining_time": "0:08:26", "throughput": 997.03, "total_tokens": 1158288} |
| {"current_steps": 118, "total_steps": 168, "loss": 0.9819, "learning_rate": 2.0308990721324927e-05, "epoch": 8.0, "percentage": 70.24, "elapsed_time": "0:19:30", "remaining_time": "0:08:16", "throughput": 997.05, "total_tokens": 1167328} |
| {"current_steps": 119, "total_steps": 168, "loss": 0.7783, "learning_rate": 1.9561928549563968e-05, "epoch": 8.067796610169491, "percentage": 70.83, "elapsed_time": "0:19:41", "remaining_time": "0:08:06", "throughput": 997.02, "total_tokens": 1177600} |
| {"current_steps": 120, "total_steps": 168, "loss": 0.7392, "learning_rate": 1.8825509907063327e-05, "epoch": 8.135593220338983, "percentage": 71.43, "elapsed_time": "0:19:49", "remaining_time": "0:07:55", "throughput": 997.23, "total_tokens": 1186256} |
| {"current_steps": 121, "total_steps": 168, "loss": 0.5762, "learning_rate": 1.8099992303349577e-05, "epoch": 8.203389830508474, "percentage": 72.02, "elapsed_time": "0:19:59", "remaining_time": "0:07:45", "throughput": 996.8, "total_tokens": 1195200} |
| {"current_steps": 122, "total_steps": 168, "loss": 0.7936, "learning_rate": 1.738562943609396e-05, "epoch": 8.271186440677965, "percentage": 72.62, "elapsed_time": "0:20:09", "remaining_time": "0:07:36", "throughput": 996.79, "total_tokens": 1205776} |
| {"current_steps": 123, "total_steps": 168, "loss": 0.794, "learning_rate": 1.6682671102399805e-05, "epoch": 8.338983050847457, "percentage": 73.21, "elapsed_time": "0:20:19", "remaining_time": "0:07:26", "throughput": 996.91, "total_tokens": 1215440} |
| {"current_steps": 124, "total_steps": 168, "loss": 0.3746, "learning_rate": 1.599136311145402e-05, "epoch": 8.40677966101695, "percentage": 73.81, "elapsed_time": "0:20:29", "remaining_time": "0:07:16", "throughput": 996.96, "total_tokens": 1226240} |
| {"current_steps": 125, "total_steps": 168, "loss": 0.5889, "learning_rate": 1.531194719857292e-05, "epoch": 8.474576271186441, "percentage": 74.4, "elapsed_time": "0:20:41", "remaining_time": "0:07:06", "throughput": 996.88, "total_tokens": 1237296} |
| {"current_steps": 126, "total_steps": 168, "loss": 0.8482, "learning_rate": 1.4644660940672627e-05, "epoch": 8.542372881355933, "percentage": 75.0, "elapsed_time": "0:20:50", "remaining_time": "0:06:56", "throughput": 997.05, "total_tokens": 1247184} |
| {"current_steps": 127, "total_steps": 168, "loss": 0.789, "learning_rate": 1.398973767319368e-05, "epoch": 8.610169491525424, "percentage": 75.6, "elapsed_time": "0:21:00", "remaining_time": "0:06:46", "throughput": 997.22, "total_tokens": 1257200} |
| {"current_steps": 128, "total_steps": 168, "loss": 0.7434, "learning_rate": 1.3347406408508695e-05, "epoch": 8.677966101694915, "percentage": 76.19, "elapsed_time": "0:21:10", "remaining_time": "0:06:37", "throughput": 997.34, "total_tokens": 1267568} |
| {"current_steps": 129, "total_steps": 168, "loss": 0.5503, "learning_rate": 1.2717891755841722e-05, "epoch": 8.745762711864407, "percentage": 76.79, "elapsed_time": "0:21:21", "remaining_time": "0:06:27", "throughput": 997.3, "total_tokens": 1278288} |
| {"current_steps": 130, "total_steps": 168, "loss": 0.5, "learning_rate": 1.2101413842727345e-05, "epoch": 8.813559322033898, "percentage": 77.38, "elapsed_time": "0:21:31", "remaining_time": "0:06:17", "throughput": 997.35, "total_tokens": 1288464} |
| {"current_steps": 131, "total_steps": 168, "loss": 0.7285, "learning_rate": 1.1498188238036861e-05, "epoch": 8.88135593220339, "percentage": 77.98, "elapsed_time": "0:21:42", "remaining_time": "0:06:07", "throughput": 996.94, "total_tokens": 1298768} |
| {"current_steps": 132, "total_steps": 168, "loss": 0.3758, "learning_rate": 1.090842587659851e-05, "epoch": 8.94915254237288, "percentage": 78.57, "elapsed_time": "0:21:52", "remaining_time": "0:05:58", "throughput": 997.0, "total_tokens": 1308912} |
| {"current_steps": 133, "total_steps": 168, "loss": 0.5785, "learning_rate": 1.0332332985438248e-05, "epoch": 9.016949152542374, "percentage": 79.17, "elapsed_time": "0:22:01", "remaining_time": "0:05:47", "throughput": 996.96, "total_tokens": 1317368} |
| {"current_steps": 134, "total_steps": 168, "loss": 0.4679, "learning_rate": 9.770111011666583e-06, "epoch": 9.084745762711865, "percentage": 79.76, "elapsed_time": "0:22:12", "remaining_time": "0:05:38", "throughput": 996.92, "total_tokens": 1328440} |
| {"current_steps": 135, "total_steps": 168, "loss": 0.6424, "learning_rate": 9.221956552036992e-06, "epoch": 9.152542372881356, "percentage": 80.36, "elapsed_time": "0:22:21", "remaining_time": "0:05:28", "throughput": 997.05, "total_tokens": 1337928} |
| {"current_steps": 136, "total_steps": 168, "loss": 0.6096, "learning_rate": 8.688061284200266e-06, "epoch": 9.220338983050848, "percentage": 80.95, "elapsed_time": "0:22:30", "remaining_time": "0:05:17", "throughput": 997.27, "total_tokens": 1346520} |
| {"current_steps": 137, "total_steps": 168, "loss": 0.5381, "learning_rate": 8.168611899679013e-06, "epoch": 9.288135593220339, "percentage": 81.55, "elapsed_time": "0:22:41", "remaining_time": "0:05:08", "throughput": 997.12, "total_tokens": 1357816} |
| {"current_steps": 138, "total_steps": 168, "loss": 0.8004, "learning_rate": 7.663790038585793e-06, "epoch": 9.35593220338983, "percentage": 82.14, "elapsed_time": "0:22:52", "remaining_time": "0:04:58", "throughput": 997.27, "total_tokens": 1368280} |
| {"current_steps": 139, "total_steps": 168, "loss": 0.4723, "learning_rate": 7.173772226107434e-06, "epoch": 9.423728813559322, "percentage": 82.74, "elapsed_time": "0:23:02", "remaining_time": "0:04:48", "throughput": 997.41, "total_tokens": 1378520} |
| {"current_steps": 140, "total_steps": 168, "loss": 0.4489, "learning_rate": 6.698729810778065e-06, "epoch": 9.491525423728813, "percentage": 83.33, "elapsed_time": "0:23:11", "remaining_time": "0:04:38", "throughput": 997.46, "total_tokens": 1388440} |
| {"current_steps": 141, "total_steps": 168, "loss": 0.707, "learning_rate": 6.238828904562316e-06, "epoch": 9.559322033898304, "percentage": 83.93, "elapsed_time": "0:23:22", "remaining_time": "0:04:28", "throughput": 996.99, "total_tokens": 1398568} |
| {"current_steps": 142, "total_steps": 168, "loss": 0.6449, "learning_rate": 5.794230324769517e-06, "epoch": 9.627118644067796, "percentage": 84.52, "elapsed_time": "0:23:33", "remaining_time": "0:04:18", "throughput": 996.84, "total_tokens": 1408856} |
| {"current_steps": 143, "total_steps": 168, "loss": 0.5634, "learning_rate": 5.365089537819434e-06, "epoch": 9.694915254237289, "percentage": 85.12, "elapsed_time": "0:23:43", "remaining_time": "0:04:08", "throughput": 996.73, "total_tokens": 1419128} |
| {"current_steps": 144, "total_steps": 168, "loss": 0.6285, "learning_rate": 4.951556604879048e-06, "epoch": 9.76271186440678, "percentage": 85.71, "elapsed_time": "0:23:54", "remaining_time": "0:03:59", "throughput": 996.63, "total_tokens": 1429944} |
| {"current_steps": 145, "total_steps": 168, "loss": 0.6617, "learning_rate": 4.5537761293894535e-06, "epoch": 9.830508474576272, "percentage": 86.31, "elapsed_time": "0:24:05", "remaining_time": "0:03:49", "throughput": 996.66, "total_tokens": 1440360} |
| {"current_steps": 146, "total_steps": 168, "loss": 0.7798, "learning_rate": 4.1718872065011904e-06, "epoch": 9.898305084745763, "percentage": 86.9, "elapsed_time": "0:24:15", "remaining_time": "0:03:39", "throughput": 996.74, "total_tokens": 1450328} |
| {"current_steps": 147, "total_steps": 168, "loss": 0.7539, "learning_rate": 3.8060233744356633e-06, "epoch": 9.966101694915254, "percentage": 87.5, "elapsed_time": "0:24:25", "remaining_time": "0:03:29", "throughput": 996.75, "total_tokens": 1461144} |
| {"current_steps": 148, "total_steps": 168, "loss": 0.3182, "learning_rate": 3.4563125677897932e-06, "epoch": 10.033898305084746, "percentage": 88.1, "elapsed_time": "0:24:34", "remaining_time": "0:03:19", "throughput": 996.66, "total_tokens": 1469864} |
| {"current_steps": 149, "total_steps": 168, "loss": 0.6796, "learning_rate": 3.1228770728000455e-06, "epoch": 10.101694915254237, "percentage": 88.69, "elapsed_time": "0:24:45", "remaining_time": "0:03:09", "throughput": 996.74, "total_tokens": 1480248} |
| {"current_steps": 150, "total_steps": 168, "loss": 0.4411, "learning_rate": 2.8058334845816213e-06, "epoch": 10.169491525423728, "percentage": 89.29, "elapsed_time": "0:24:52", "remaining_time": "0:02:59", "throughput": 996.86, "total_tokens": 1488056} |
| {"current_steps": 151, "total_steps": 168, "loss": 0.6203, "learning_rate": 2.5052926663577e-06, "epoch": 10.23728813559322, "percentage": 89.88, "elapsed_time": "0:25:02", "remaining_time": "0:02:49", "throughput": 996.54, "total_tokens": 1496920} |
| {"current_steps": 152, "total_steps": 168, "loss": 0.6025, "learning_rate": 2.221359710692961e-06, "epoch": 10.305084745762711, "percentage": 90.48, "elapsed_time": "0:25:12", "remaining_time": "0:02:39", "throughput": 996.57, "total_tokens": 1507144} |
| {"current_steps": 153, "total_steps": 168, "loss": 0.5943, "learning_rate": 1.9541339027450256e-06, "epoch": 10.372881355932204, "percentage": 91.07, "elapsed_time": "0:25:20", "remaining_time": "0:02:29", "throughput": 996.79, "total_tokens": 1515560} |
| {"current_steps": 154, "total_steps": 168, "loss": 0.5806, "learning_rate": 1.70370868554659e-06, "epoch": 10.440677966101696, "percentage": 91.67, "elapsed_time": "0:25:31", "remaining_time": "0:02:19", "throughput": 996.81, "total_tokens": 1526296} |
| {"current_steps": 155, "total_steps": 168, "loss": 0.7033, "learning_rate": 1.4701716273304521e-06, "epoch": 10.508474576271187, "percentage": 92.26, "elapsed_time": "0:25:42", "remaining_time": "0:02:09", "throughput": 996.78, "total_tokens": 1537064} |
| {"current_steps": 156, "total_steps": 168, "loss": 0.629, "learning_rate": 1.2536043909088191e-06, "epoch": 10.576271186440678, "percentage": 92.86, "elapsed_time": "0:25:52", "remaining_time": "0:01:59", "throughput": 996.85, "total_tokens": 1547816} |
| {"current_steps": 157, "total_steps": 168, "loss": 0.6927, "learning_rate": 1.0540827051175818e-06, "epoch": 10.64406779661017, "percentage": 93.45, "elapsed_time": "0:26:02", "remaining_time": "0:01:49", "throughput": 996.89, "total_tokens": 1557672} |
| {"current_steps": 158, "total_steps": 168, "loss": 0.4882, "learning_rate": 8.716763383355864e-07, "epoch": 10.711864406779661, "percentage": 94.05, "elapsed_time": "0:26:13", "remaining_time": "0:01:39", "throughput": 996.93, "total_tokens": 1568264} |
| {"current_steps": 159, "total_steps": 168, "loss": 0.543, "learning_rate": 7.064490740882057e-07, "epoch": 10.779661016949152, "percentage": 94.64, "elapsed_time": "0:26:23", "remaining_time": "0:01:29", "throughput": 996.93, "total_tokens": 1578744} |
| {"current_steps": 160, "total_steps": 168, "loss": 0.3713, "learning_rate": 5.584586887435739e-07, "epoch": 10.847457627118644, "percentage": 95.24, "elapsed_time": "0:26:32", "remaining_time": "0:01:19", "throughput": 996.96, "total_tokens": 1588008} |
| {"current_steps": 161, "total_steps": 168, "loss": 0.5766, "learning_rate": 4.277569313094809e-07, "epoch": 10.915254237288135, "percentage": 95.83, "elapsed_time": "0:26:41", "remaining_time": "0:01:09", "throughput": 996.69, "total_tokens": 1596200} |
| {"current_steps": 162, "total_steps": 168, "loss": 0.5419, "learning_rate": 3.143895053378698e-07, "epoch": 10.983050847457626, "percentage": 96.43, "elapsed_time": "0:26:52", "remaining_time": "0:00:59", "throughput": 996.72, "total_tokens": 1606760} |
| {"current_steps": 163, "total_steps": 168, "loss": 0.5242, "learning_rate": 2.1839605294330933e-07, "epoch": 11.05084745762712, "percentage": 97.02, "elapsed_time": "0:26:59", "remaining_time": "0:00:49", "throughput": 996.87, "total_tokens": 1614864} |
| {"current_steps": 164, "total_steps": 168, "loss": 0.4455, "learning_rate": 1.3981014094099353e-07, "epoch": 11.11864406779661, "percentage": 97.62, "elapsed_time": "0:27:08", "remaining_time": "0:00:39", "throughput": 996.99, "total_tokens": 1624048} |
| {"current_steps": 165, "total_steps": 168, "loss": 0.6553, "learning_rate": 7.865924910916977e-08, "epoch": 11.186440677966102, "percentage": 98.21, "elapsed_time": "0:27:18", "remaining_time": "0:00:29", "throughput": 997.11, "total_tokens": 1633808} |
| {"current_steps": 166, "total_steps": 168, "loss": 0.4241, "learning_rate": 3.496476058006959e-08, "epoch": 11.254237288135593, "percentage": 98.81, "elapsed_time": "0:27:29", "remaining_time": "0:00:19", "throughput": 997.04, "total_tokens": 1644560} |
| {"current_steps": 167, "total_steps": 168, "loss": 0.7591, "learning_rate": 8.741954362678772e-09, "epoch": 11.322033898305085, "percentage": 99.4, "elapsed_time": "0:27:40", "remaining_time": "0:00:09", "throughput": 997.05, "total_tokens": 1655392} |
| {"current_steps": 168, "total_steps": 168, "loss": 0.5966, "learning_rate": 0.0, "epoch": 11.389830508474576, "percentage": 100.0, "elapsed_time": "0:27:49", "remaining_time": "0:00:00", "throughput": 997.08, "total_tokens": 1664720} |
| {"current_steps": 168, "total_steps": 168, "epoch": 11.389830508474576, "percentage": 100.0, "elapsed_time": "0:27:50", "remaining_time": "0:00:00", "throughput": 996.65, "total_tokens": 1664720} |
|
|