|
|
{"loss": 2.0249629, "token_acc": 0.59259259, "grad_norm": 49.5, "learning_rate": 2e-07, "memory(GiB)": 30.5, "train_speed(iter/s)": 0.096714, "epoch": 0.00207819, "global_step/max_steps": "1/481", "percentage": "0.21%", "elapsed_time": "10s", "remaining_time": "1h 20m 52s"} |
|
|
{"loss": 1.91072392, "token_acc": 0.60309682, "grad_norm": 48.25, "learning_rate": 1e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.104169, "epoch": 0.01039096, "global_step/max_steps": "5/481", "percentage": "1.04%", "elapsed_time": "47s", "remaining_time": "1h 15m 47s"} |
|
|
{"loss": 1.88561249, "token_acc": 0.60607252, "grad_norm": 46.0, "learning_rate": 2e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.10431, "epoch": 0.02078192, "global_step/max_steps": "10/481", "percentage": "2.08%", "elapsed_time": "1m 35s", "remaining_time": "1h 15m 4s"} |
|
|
{"loss": 1.66712418, "token_acc": 0.62767596, "grad_norm": 41.25, "learning_rate": 3e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.104357, "epoch": 0.03117288, "global_step/max_steps": "15/481", "percentage": "3.12%", "elapsed_time": "2m 23s", "remaining_time": "1h 14m 18s"} |
|
|
{"loss": 1.42836103, "token_acc": 0.64458779, "grad_norm": 35.75, "learning_rate": 4e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.104096, "epoch": 0.04156384, "global_step/max_steps": "20/481", "percentage": "4.16%", "elapsed_time": "3m 11s", "remaining_time": "1h 13m 43s"} |
|
|
{"loss": 0.85813541, "token_acc": 0.73619441, "grad_norm": 24.125, "learning_rate": 5e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103919, "epoch": 0.0519548, "global_step/max_steps": "25/481", "percentage": "5.20%", "elapsed_time": "4m 0s", "remaining_time": "1h 13m 3s"} |
|
|
{"loss": 0.39419129, "token_acc": 0.8509957, "grad_norm": 10.4375, "learning_rate": 5e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103913, "epoch": 0.06234576, "global_step/max_steps": "30/481", "percentage": "6.24%", "elapsed_time": "4m 48s", "remaining_time": "1h 12m 16s"} |
|
|
{"loss": 0.24166102, "token_acc": 0.91070042, "grad_norm": 3.6875, "learning_rate": 4.99e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103983, "epoch": 0.07273672, "global_step/max_steps": "35/481", "percentage": "7.28%", "elapsed_time": "5m 36s", "remaining_time": "1h 11m 26s"} |
|
|
{"loss": 0.20080657, "token_acc": 0.92055241, "grad_norm": 1.828125, "learning_rate": 4.99e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103895, "epoch": 0.08312768, "global_step/max_steps": "40/481", "percentage": "8.32%", "elapsed_time": "6m 24s", "remaining_time": "1h 10m 42s"} |
|
|
{"loss": 0.19015437, "token_acc": 0.91806477, "grad_norm": 2.265625, "learning_rate": 4.98e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103752, "epoch": 0.09351864, "global_step/max_steps": "45/481", "percentage": "9.36%", "elapsed_time": "7m 13s", "remaining_time": "1h 10m 0s"} |
|
|
{"loss": 0.19012609, "token_acc": 0.9183419, "grad_norm": 1.9921875, "learning_rate": 4.96e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103736, "epoch": 0.1039096, "global_step/max_steps": "50/481", "percentage": "10.40%", "elapsed_time": "8m 1s", "remaining_time": "1h 9m 12s"} |
|
|
{"loss": 0.18768159, "token_acc": 0.91742037, "grad_norm": 1.9921875, "learning_rate": 4.95e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103693, "epoch": 0.11430056, "global_step/max_steps": "55/481", "percentage": "11.43%", "elapsed_time": "8m 50s", "remaining_time": "1h 8m 26s"} |
|
|
{"loss": 0.18475006, "token_acc": 0.91948694, "grad_norm": 1.640625, "learning_rate": 4.93e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103749, "epoch": 0.12469152, "global_step/max_steps": "60/481", "percentage": "12.47%", "elapsed_time": "9m 38s", "remaining_time": "1h 7m 36s"} |
|
|
{"loss": 0.18700193, "token_acc": 0.91834238, "grad_norm": 2.03125, "learning_rate": 4.91e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103789, "epoch": 0.13508248, "global_step/max_steps": "65/481", "percentage": "13.51%", "elapsed_time": "10m 26s", "remaining_time": "1h 6m 46s"} |
|
|
{"loss": 0.18332734, "token_acc": 0.9198804, "grad_norm": 1.8984375, "learning_rate": 4.88e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103777, "epoch": 0.14547344, "global_step/max_steps": "70/481", "percentage": "14.55%", "elapsed_time": "11m 14s", "remaining_time": "1h 5m 59s"} |
|
|
{"loss": 0.17667145, "token_acc": 0.92212375, "grad_norm": 1.8046875, "learning_rate": 4.85e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.101758, "epoch": 0.1558644, "global_step/max_steps": "75/481", "percentage": "15.59%", "elapsed_time": "12m 16s", "remaining_time": "1h 6m 28s"} |
|
|
{"loss": 0.17785825, "token_acc": 0.92087503, "grad_norm": 1.8203125, "learning_rate": 4.82e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.096873, "epoch": 0.16625536, "global_step/max_steps": "80/481", "percentage": "16.63%", "elapsed_time": "13m 45s", "remaining_time": "1h 8m 58s"} |
|
|
{"loss": 0.18061868, "token_acc": 0.92047299, "grad_norm": 1.5390625, "learning_rate": 4.79e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.092995, "epoch": 0.17664632, "global_step/max_steps": "85/481", "percentage": "17.67%", "elapsed_time": "15m 13s", "remaining_time": "1h 10m 57s"} |
|
|
{"loss": 0.17701297, "token_acc": 0.92251349, "grad_norm": 1.8828125, "learning_rate": 4.75e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.089657, "epoch": 0.18703728, "global_step/max_steps": "90/481", "percentage": "18.71%", "elapsed_time": "16m 43s", "remaining_time": "1h 12m 40s"} |
|
|
{"loss": 0.18493406, "token_acc": 0.91693266, "grad_norm": 3.21875, "learning_rate": 4.71e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.086893, "epoch": 0.19742824, "global_step/max_steps": "95/481", "percentage": "19.75%", "elapsed_time": "18m 13s", "remaining_time": "1h 14m 1s"} |
|
|
{"loss": 0.1811494, "token_acc": 0.91940932, "grad_norm": 3.3125, "learning_rate": 4.67e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.084511, "epoch": 0.2078192, "global_step/max_steps": "100/481", "percentage": "20.79%", "elapsed_time": "19m 43s", "remaining_time": "1h 15m 7s"} |
|
|
{"eval_loss": 0.17230889, "eval_token_acc": 0.92528807, "eval_runtime": 1320.3184, "eval_samples_per_second": 60.881, "eval_steps_per_second": 3.805, "epoch": 0.2078192, "global_step/max_steps": "100/481", "percentage": "20.79%", "elapsed_time": "41m 43s", "remaining_time": "2h 38m 57s"} |
|
|
{"loss": 0.17557873, "token_acc": 0.92525432, "grad_norm": 1.9375, "learning_rate": 4.63e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03748, "epoch": 0.21821016, "global_step/max_steps": "105/481", "percentage": "21.83%", "elapsed_time": "46m 41s", "remaining_time": "2h 47m 11s"} |
|
|
{"loss": 0.16902719, "token_acc": 0.92608154, "grad_norm": 1.6484375, "learning_rate": 4.58e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.0381, "epoch": 0.22860112, "global_step/max_steps": "110/481", "percentage": "22.87%", "elapsed_time": "48m 6s", "remaining_time": "2h 42m 16s"} |
|
|
{"loss": 0.17990721, "token_acc": 0.91966157, "grad_norm": 3.0, "learning_rate": 4.53e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038616, "epoch": 0.23899208, "global_step/max_steps": "115/481", "percentage": "23.91%", "elapsed_time": "49m 37s", "remaining_time": "2h 37m 57s"} |
|
|
{"loss": 0.17740324, "token_acc": 0.92122452, "grad_norm": 1.984375, "learning_rate": 4.48e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039106, "epoch": 0.24938304, "global_step/max_steps": "120/481", "percentage": "24.95%", "elapsed_time": "51m 8s", "remaining_time": "2h 33m 50s"} |
|
|
{"loss": 0.17831547, "token_acc": 0.91834229, "grad_norm": 2.03125, "learning_rate": 4.43e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039569, "epoch": 0.259774, "global_step/max_steps": "125/481", "percentage": "25.99%", "elapsed_time": "52m 38s", "remaining_time": "2h 29m 56s"} |
|
|
{"loss": 0.17632855, "token_acc": 0.92014238, "grad_norm": 1.5546875, "learning_rate": 4.37e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.040384, "epoch": 0.27016496, "global_step/max_steps": "130/481", "percentage": "27.03%", "elapsed_time": "53m 39s", "remaining_time": "2h 24m 53s"} |
|
|
{"loss": 0.17515565, "token_acc": 0.92026369, "grad_norm": 1.625, "learning_rate": 4.32e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.041134, "epoch": 0.28055592, "global_step/max_steps": "135/481", "percentage": "28.07%", "elapsed_time": "54m 42s", "remaining_time": "2h 20m 12s"} |
|
|
{"loss": 0.18066968, "token_acc": 0.91946882, "grad_norm": 2.09375, "learning_rate": 4.26e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.041899, "epoch": 0.29094688, "global_step/max_steps": "140/481", "percentage": "29.11%", "elapsed_time": "55m 41s", "remaining_time": "2h 15m 39s"} |
|
|
{"loss": 0.18014355, "token_acc": 0.91789935, "grad_norm": 1.140625, "learning_rate": 4.19e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.042363, "epoch": 0.30133784, "global_step/max_steps": "145/481", "percentage": "30.15%", "elapsed_time": "57m 2s", "remaining_time": "2h 12m 10s"} |
|
|
{"loss": 0.17886133, "token_acc": 0.92107931, "grad_norm": 1.3515625, "learning_rate": 4.13e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.042968, "epoch": 0.3117288, "global_step/max_steps": "150/481", "percentage": "31.19%", "elapsed_time": "58m 10s", "remaining_time": "2h 8m 22s"} |
|
|
{"loss": 0.18089871, "token_acc": 0.91918955, "grad_norm": 1.7734375, "learning_rate": 4.06e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.043361, "epoch": 0.32211976, "global_step/max_steps": "155/481", "percentage": "32.22%", "elapsed_time": "59m 34s", "remaining_time": "2h 5m 18s"} |
|
|
{"loss": 0.17732292, "token_acc": 0.91881111, "grad_norm": 1.7421875, "learning_rate": 3.99e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.043692, "epoch": 0.33251072, "global_step/max_steps": "160/481", "percentage": "33.26%", "elapsed_time": "1h 1m 1s", "remaining_time": "2h 2m 26s"} |
|
|
{"loss": 0.17417793, "token_acc": 0.92164034, "grad_norm": 1.78125, "learning_rate": 3.92e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.043984, "epoch": 0.34290168, "global_step/max_steps": "165/481", "percentage": "34.30%", "elapsed_time": "1h 2m 31s", "remaining_time": "1h 59m 44s"} |
|
|
{"loss": 0.17619532, "token_acc": 0.92279668, "grad_norm": 1.875, "learning_rate": 3.85e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.044279, "epoch": 0.35329264, "global_step/max_steps": "170/481", "percentage": "35.34%", "elapsed_time": "1h 3m 59s", "remaining_time": "1h 57m 3s"} |
|
|
{"loss": 0.1743765, "token_acc": 0.92309516, "grad_norm": 3.71875, "learning_rate": 3.78e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.044551, "epoch": 0.3636836, "global_step/max_steps": "175/481", "percentage": "36.38%", "elapsed_time": "1h 5m 27s", "remaining_time": "1h 54m 28s"} |
|
|
{"loss": 0.17724402, "token_acc": 0.92029781, "grad_norm": 1.84375, "learning_rate": 3.7e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.044808, "epoch": 0.37407456, "global_step/max_steps": "180/481", "percentage": "37.42%", "elapsed_time": "1h 6m 56s", "remaining_time": "1h 51m 57s"} |
|
|
{"loss": 0.18116597, "token_acc": 0.91708861, "grad_norm": 2.1875, "learning_rate": 3.63e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.045056, "epoch": 0.38446552, "global_step/max_steps": "185/481", "percentage": "38.46%", "elapsed_time": "1h 8m 25s", "remaining_time": "1h 49m 29s"} |
|
|
{"loss": 0.17593677, "token_acc": 0.92025862, "grad_norm": 2.046875, "learning_rate": 3.55e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.045325, "epoch": 0.39485647, "global_step/max_steps": "190/481", "percentage": "39.50%", "elapsed_time": "1h 9m 52s", "remaining_time": "1h 47m 1s"} |
|
|
{"loss": 0.182314, "token_acc": 0.91763968, "grad_norm": 2.0, "learning_rate": 3.47e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.04554, "epoch": 0.40524743, "global_step/max_steps": "195/481", "percentage": "40.54%", "elapsed_time": "1h 11m 21s", "remaining_time": "1h 44m 39s"} |
|
|
{"loss": 0.18079571, "token_acc": 0.91806215, "grad_norm": 1.7109375, "learning_rate": 3.39e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.045874, "epoch": 0.41563839, "global_step/max_steps": "200/481", "percentage": "41.58%", "elapsed_time": "1h 12m 39s", "remaining_time": "1h 42m 5s"} |
|
|
{"eval_loss": 0.17064263, "eval_token_acc": 0.92389875, "eval_runtime": 1310.4171, "eval_samples_per_second": 61.341, "eval_steps_per_second": 3.834, "epoch": 0.41563839, "global_step/max_steps": "200/481", "percentage": "41.58%", "elapsed_time": "1h 34m 30s", "remaining_time": "2h 12m 46s"} |
|
|
{"loss": 0.18160745, "token_acc": 0.92382996, "grad_norm": 1.7890625, "learning_rate": 3.31e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034354, "epoch": 0.42602935, "global_step/max_steps": "205/481", "percentage": "42.62%", "elapsed_time": "1h 39m 27s", "remaining_time": "2h 13m 53s"} |
|
|
{"loss": 0.1788413, "token_acc": 0.91801604, "grad_norm": 2.0625, "learning_rate": 3.23e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034671, "epoch": 0.43642031, "global_step/max_steps": "210/481", "percentage": "43.66%", "elapsed_time": "1h 40m 56s", "remaining_time": "2h 10m 15s"} |
|
|
{"loss": 0.17460221, "token_acc": 0.92020492, "grad_norm": 1.484375, "learning_rate": 3.15e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034976, "epoch": 0.44681127, "global_step/max_steps": "215/481", "percentage": "44.70%", "elapsed_time": "1h 42m 26s", "remaining_time": "2h 6m 45s"} |
|
|
{"loss": 0.17835221, "token_acc": 0.91984046, "grad_norm": 1.234375, "learning_rate": 3.06e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035377, "epoch": 0.45720223, "global_step/max_steps": "220/481", "percentage": "45.74%", "elapsed_time": "1h 43m 39s", "remaining_time": "2h 2m 58s"} |
|
|
{"loss": 0.17963867, "token_acc": 0.92075208, "grad_norm": 1.78125, "learning_rate": 2.98e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035838, "epoch": 0.46759319, "global_step/max_steps": "225/481", "percentage": "46.78%", "elapsed_time": "1h 44m 37s", "remaining_time": "1h 59m 2s"} |
|
|
{"loss": 0.18421537, "token_acc": 0.91660908, "grad_norm": 1.921875, "learning_rate": 2.89e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036335, "epoch": 0.47798415, "global_step/max_steps": "230/481", "percentage": "47.82%", "elapsed_time": "1h 45m 30s", "remaining_time": "1h 55m 8s"} |
|
|
{"loss": 0.17683053, "token_acc": 0.92130919, "grad_norm": 1.7890625, "learning_rate": 2.81e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036825, "epoch": 0.48837511, "global_step/max_steps": "235/481", "percentage": "48.86%", "elapsed_time": "1h 46m 21s", "remaining_time": "1h 51m 20s"} |
|
|
{"loss": 0.17713736, "token_acc": 0.92035536, "grad_norm": 1.84375, "learning_rate": 2.72e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037143, "epoch": 0.49876607, "global_step/max_steps": "240/481", "percentage": "49.90%", "elapsed_time": "1h 47m 41s", "remaining_time": "1h 48m 8s"} |
|
|
{"loss": 0.17735506, "token_acc": 0.9191974, "grad_norm": 1.5234375, "learning_rate": 2.64e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037405, "epoch": 0.50915703, "global_step/max_steps": "245/481", "percentage": "50.94%", "elapsed_time": "1h 49m 9s", "remaining_time": "1h 45m 9s"} |
|
|
{"loss": 0.18297757, "token_acc": 0.91939188, "grad_norm": 1.3828125, "learning_rate": 2.55e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037652, "epoch": 0.51954799, "global_step/max_steps": "250/481", "percentage": "51.98%", "elapsed_time": "1h 50m 39s", "remaining_time": "1h 42m 14s"} |
|
|
{"loss": 0.17588781, "token_acc": 0.92162738, "grad_norm": 2.203125, "learning_rate": 2.47e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037896, "epoch": 0.52993895, "global_step/max_steps": "255/481", "percentage": "53.01%", "elapsed_time": "1h 52m 8s", "remaining_time": "1h 39m 23s"} |
|
|
{"loss": 0.18080204, "token_acc": 0.92120322, "grad_norm": 1.6796875, "learning_rate": 2.38e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038125, "epoch": 0.54032991, "global_step/max_steps": "260/481", "percentage": "54.05%", "elapsed_time": "1h 53m 39s", "remaining_time": "1h 36m 36s"} |
|
|
{"loss": 0.18104695, "token_acc": 0.91801708, "grad_norm": 2.6875, "learning_rate": 2.29e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038349, "epoch": 0.55072087, "global_step/max_steps": "265/481", "percentage": "55.09%", "elapsed_time": "1h 55m 9s", "remaining_time": "1h 33m 52s"} |
|
|
{"loss": 0.17682294, "token_acc": 0.92037196, "grad_norm": 1.7578125, "learning_rate": 2.21e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038567, "epoch": 0.56111183, "global_step/max_steps": "270/481", "percentage": "56.13%", "elapsed_time": "1h 56m 40s", "remaining_time": "1h 31m 10s"} |
|
|
{"loss": 0.17726024, "token_acc": 0.91912933, "grad_norm": 1.34375, "learning_rate": 2.12e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038795, "epoch": 0.57150279, "global_step/max_steps": "275/481", "percentage": "57.17%", "elapsed_time": "1h 58m 8s", "remaining_time": "1h 28m 29s"} |
|
|
{"loss": 0.18150538, "token_acc": 0.91745613, "grad_norm": 1.71875, "learning_rate": 2.04e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03901, "epoch": 0.58189375, "global_step/max_steps": "280/481", "percentage": "58.21%", "elapsed_time": "1h 59m 37s", "remaining_time": "1h 25m 52s"} |
|
|
{"loss": 0.17500284, "token_acc": 0.92282113, "grad_norm": 1.546875, "learning_rate": 1.95e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039295, "epoch": 0.59228471, "global_step/max_steps": "285/481", "percentage": "59.25%", "elapsed_time": "2h 0m 52s", "remaining_time": "1h 23m 7s"} |
|
|
{"loss": 0.17536248, "token_acc": 0.92083558, "grad_norm": 1.234375, "learning_rate": 1.87e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039587, "epoch": 0.60267567, "global_step/max_steps": "290/481", "percentage": "60.29%", "elapsed_time": "2h 2m 5s", "remaining_time": "1h 20m 24s"} |
|
|
{"loss": 0.17674764, "token_acc": 0.92204176, "grad_norm": 1.5390625, "learning_rate": 1.79e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039803, "epoch": 0.61306663, "global_step/max_steps": "295/481", "percentage": "61.33%", "elapsed_time": "2h 3m 31s", "remaining_time": "1h 17m 52s"} |
|
|
{"loss": 0.17351619, "token_acc": 0.92283452, "grad_norm": 1.9140625, "learning_rate": 1.7e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.040008, "epoch": 0.62345759, "global_step/max_steps": "300/481", "percentage": "62.37%", "elapsed_time": "2h 4m 58s", "remaining_time": "1h 15m 24s"} |
|
|
{"eval_loss": 0.16996929, "eval_token_acc": 0.92522779, "eval_runtime": 1338.6589, "eval_samples_per_second": 60.047, "eval_steps_per_second": 3.753, "epoch": 0.62345759, "global_step/max_steps": "300/481", "percentage": "62.37%", "elapsed_time": "2h 27m 17s", "remaining_time": "1h 28m 51s"} |
|
|
{"loss": 0.17297822, "token_acc": 0.92521466, "grad_norm": 1.7109375, "learning_rate": 1.62e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033422, "epoch": 0.63384855, "global_step/max_steps": "305/481", "percentage": "63.41%", "elapsed_time": "2h 32m 5s", "remaining_time": "1h 27m 45s"} |
|
|
{"loss": 0.17585882, "token_acc": 0.92024102, "grad_norm": 2.296875, "learning_rate": 1.54e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033685, "epoch": 0.64423951, "global_step/max_steps": "310/481", "percentage": "64.45%", "elapsed_time": "2h 33m 22s", "remaining_time": "1h 24m 36s"} |
|
|
{"loss": 0.17637254, "token_acc": 0.92029497, "grad_norm": 2.03125, "learning_rate": 1.46e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034017, "epoch": 0.65463047, "global_step/max_steps": "315/481", "percentage": "65.49%", "elapsed_time": "2h 34m 20s", "remaining_time": "1h 21m 20s"} |
|
|
{"loss": 0.18037752, "token_acc": 0.91979536, "grad_norm": 2.296875, "learning_rate": 1.39e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034317, "epoch": 0.66502143, "global_step/max_steps": "320/481", "percentage": "66.53%", "elapsed_time": "2h 35m 25s", "remaining_time": "1h 18m 11s"} |
|
|
{"loss": 0.17441235, "token_acc": 0.92338146, "grad_norm": 2.046875, "learning_rate": 1.31e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034642, "epoch": 0.67541239, "global_step/max_steps": "325/481", "percentage": "67.57%", "elapsed_time": "2h 36m 22s", "remaining_time": "1h 15m 3s"} |
|
|
{"loss": 0.17859527, "token_acc": 0.92036657, "grad_norm": 1.640625, "learning_rate": 1.24e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034906, "epoch": 0.68580335, "global_step/max_steps": "330/481", "percentage": "68.61%", "elapsed_time": "2h 37m 33s", "remaining_time": "1h 12m 5s"} |
|
|
{"loss": 0.17650687, "token_acc": 0.9200735, "grad_norm": 1.375, "learning_rate": 1.16e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035109, "epoch": 0.69619431, "global_step/max_steps": "335/481", "percentage": "69.65%", "elapsed_time": "2h 39m 1s", "remaining_time": "1h 9m 18s"} |
|
|
{"loss": 0.18125291, "token_acc": 0.9172004, "grad_norm": 1.7734375, "learning_rate": 1.09e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035306, "epoch": 0.70658527, "global_step/max_steps": "340/481", "percentage": "70.69%", "elapsed_time": "2h 40m 29s", "remaining_time": "1h 6m 33s"} |
|
|
{"loss": 0.18101528, "token_acc": 0.91588857, "grad_norm": 1.046875, "learning_rate": 1.02e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035495, "epoch": 0.71697623, "global_step/max_steps": "345/481", "percentage": "71.73%", "elapsed_time": "2h 41m 59s", "remaining_time": "1h 3m 51s"} |
|
|
{"loss": 0.17332951, "token_acc": 0.92229359, "grad_norm": 2.203125, "learning_rate": 9.5e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035686, "epoch": 0.72736719, "global_step/max_steps": "350/481", "percentage": "72.77%", "elapsed_time": "2h 43m 27s", "remaining_time": "1h 1m 10s"} |
|
|
{"loss": 0.17291888, "token_acc": 0.9233902, "grad_norm": 1.28125, "learning_rate": 8.8e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035869, "epoch": 0.73775815, "global_step/max_steps": "355/481", "percentage": "73.80%", "elapsed_time": "2h 44m 56s", "remaining_time": "58m 32s"} |
|
|
{"loss": 0.17527255, "token_acc": 0.91944069, "grad_norm": 1.5859375, "learning_rate": 8.2e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03605, "epoch": 0.74814911, "global_step/max_steps": "360/481", "percentage": "74.84%", "elapsed_time": "2h 46m 25s", "remaining_time": "55m 56s"} |
|
|
{"loss": 0.17583263, "token_acc": 0.92228378, "grad_norm": 1.8984375, "learning_rate": 7.6e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036228, "epoch": 0.75854007, "global_step/max_steps": "365/481", "percentage": "75.88%", "elapsed_time": "2h 47m 54s", "remaining_time": "53m 21s"} |
|
|
{"loss": 0.17296208, "token_acc": 0.91996611, "grad_norm": 1.4453125, "learning_rate": 7e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036459, "epoch": 0.76893103, "global_step/max_steps": "370/481", "percentage": "76.92%", "elapsed_time": "2h 49m 8s", "remaining_time": "50m 44s"} |
|
|
{"loss": 0.17531843, "token_acc": 0.92152662, "grad_norm": 2.15625, "learning_rate": 6.4e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036681, "epoch": 0.77932199, "global_step/max_steps": "375/481", "percentage": "77.96%", "elapsed_time": "2h 50m 23s", "remaining_time": "48m 9s"} |
|
|
{"loss": 0.17355837, "token_acc": 0.92119117, "grad_norm": 1.453125, "learning_rate": 5.8e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036859, "epoch": 0.78971295, "global_step/max_steps": "380/481", "percentage": "79.00%", "elapsed_time": "2h 51m 49s", "remaining_time": "45m 40s"} |
|
|
{"loss": 0.18347428, "token_acc": 0.91669904, "grad_norm": 2.03125, "learning_rate": 5.3e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037022, "epoch": 0.80010391, "global_step/max_steps": "385/481", "percentage": "80.04%", "elapsed_time": "2h 53m 18s", "remaining_time": "43m 12s"} |
|
|
{"loss": 0.17866383, "token_acc": 0.91900555, "grad_norm": 1.78125, "learning_rate": 4.8e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037179, "epoch": 0.81049487, "global_step/max_steps": "390/481", "percentage": "81.08%", "elapsed_time": "2h 54m 49s", "remaining_time": "40m 47s"} |
|
|
{"loss": 0.16837071, "token_acc": 0.92525205, "grad_norm": 1.7265625, "learning_rate": 4.3e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037337, "epoch": 0.82088583, "global_step/max_steps": "395/481", "percentage": "82.12%", "elapsed_time": "2h 56m 19s", "remaining_time": "38m 23s"} |
|
|
{"loss": 0.17911137, "token_acc": 0.9203882, "grad_norm": 1.625, "learning_rate": 3.8e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037494, "epoch": 0.83127679, "global_step/max_steps": "400/481", "percentage": "83.16%", "elapsed_time": "2h 57m 48s", "remaining_time": "36m 0s"} |
|
|
{"eval_loss": 0.16972463, "eval_token_acc": 0.92390072, "eval_runtime": 1325.8316, "eval_samples_per_second": 60.628, "eval_steps_per_second": 3.789, "epoch": 0.83127679, "global_step/max_steps": "400/481", "percentage": "83.16%", "elapsed_time": "3h 19m 54s", "remaining_time": "40m 28s"} |
|
|
{"loss": 0.1712463, "token_acc": 0.92386819, "grad_norm": 1.546875, "learning_rate": 3.3e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.032973, "epoch": 0.84166775, "global_step/max_steps": "405/481", "percentage": "84.20%", "elapsed_time": "3h 24m 43s", "remaining_time": "38m 25s"} |
|
|
{"loss": 0.17123722, "token_acc": 0.92165175, "grad_norm": 2.25, "learning_rate": 2.9e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033214, "epoch": 0.85205871, "global_step/max_steps": "410/481", "percentage": "85.24%", "elapsed_time": "3h 25m 43s", "remaining_time": "35m 37s"} |
|
|
{"loss": 0.17746515, "token_acc": 0.91808238, "grad_norm": 1.5859375, "learning_rate": 2.5e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033458, "epoch": 0.86244967, "global_step/max_steps": "415/481", "percentage": "86.28%", "elapsed_time": "3h 26m 44s", "remaining_time": "32m 52s"} |
|
|
{"loss": 0.17772366, "token_acc": 0.91908013, "grad_norm": 1.3984375, "learning_rate": 2.2e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033705, "epoch": 0.87284063, "global_step/max_steps": "420/481", "percentage": "87.32%", "elapsed_time": "3h 27m 41s", "remaining_time": "30m 9s"} |
|
|
{"loss": 0.17560486, "token_acc": 0.9201813, "grad_norm": 1.53125, "learning_rate": 1.8e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033868, "epoch": 0.88323159, "global_step/max_steps": "425/481", "percentage": "88.36%", "elapsed_time": "3h 29m 8s", "remaining_time": "27m 33s"} |
|
|
{"loss": 0.17915328, "token_acc": 0.91824886, "grad_norm": 1.7578125, "learning_rate": 1.5e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034029, "epoch": 0.89362255, "global_step/max_steps": "430/481", "percentage": "89.40%", "elapsed_time": "3h 30m 36s", "remaining_time": "24m 58s"} |
|
|
{"loss": 0.17606736, "token_acc": 0.92286885, "grad_norm": 1.3671875, "learning_rate": 1.2e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034183, "epoch": 0.90401351, "global_step/max_steps": "435/481", "percentage": "90.44%", "elapsed_time": "3h 32m 5s", "remaining_time": "22m 25s"} |
|
|
{"loss": 0.17585526, "token_acc": 0.91838664, "grad_norm": 1.546875, "learning_rate": 1e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034333, "epoch": 0.91440447, "global_step/max_steps": "440/481", "percentage": "91.48%", "elapsed_time": "3h 33m 35s", "remaining_time": "19m 54s"} |
|
|
{"loss": 0.16930281, "token_acc": 0.92383628, "grad_norm": 1.5, "learning_rate": 8e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034485, "epoch": 0.92479543, "global_step/max_steps": "445/481", "percentage": "92.52%", "elapsed_time": "3h 35m 3s", "remaining_time": "17m 23s"} |
|
|
{"loss": 0.17058501, "token_acc": 0.925625, "grad_norm": 2.078125, "learning_rate": 6e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03464, "epoch": 0.93518639, "global_step/max_steps": "450/481", "percentage": "93.56%", "elapsed_time": "3h 36m 30s", "remaining_time": "14m 54s"} |
|
|
{"loss": 0.18178566, "token_acc": 0.92082941, "grad_norm": 1.5703125, "learning_rate": 4e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03485, "epoch": 0.94557735, "global_step/max_steps": "455/481", "percentage": "94.59%", "elapsed_time": "3h 37m 35s", "remaining_time": "12m 26s"} |
|
|
{"loss": 0.17831986, "token_acc": 0.92116054, "grad_norm": 1.234375, "learning_rate": 3e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035001, "epoch": 0.95596831, "global_step/max_steps": "460/481", "percentage": "95.63%", "elapsed_time": "3h 39m 2s", "remaining_time": "9m 59s"} |
|
|
{"loss": 0.1767773, "token_acc": 0.91808686, "grad_norm": 1.59375, "learning_rate": 2e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03514, "epoch": 0.96635927, "global_step/max_steps": "465/481", "percentage": "96.67%", "elapsed_time": "3h 40m 32s", "remaining_time": "7m 35s"} |
|
|
{"loss": 0.17313094, "token_acc": 0.92246365, "grad_norm": 1.984375, "learning_rate": 1e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03528, "epoch": 0.97675023, "global_step/max_steps": "470/481", "percentage": "97.71%", "elapsed_time": "3h 42m 1s", "remaining_time": "5m 11s"} |
|
|
{"loss": 0.17898101, "token_acc": 0.91830571, "grad_norm": 2.0625, "learning_rate": 0.0, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035421, "epoch": 0.98714119, "global_step/max_steps": "475/481", "percentage": "98.75%", "elapsed_time": "3h 43m 29s", "remaining_time": "2m 49s"} |
|
|
{"loss": 0.17718053, "token_acc": 0.91961875, "grad_norm": 1.296875, "learning_rate": 0.0, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03556, "epoch": 0.99753215, "global_step/max_steps": "480/481", "percentage": "99.79%", "elapsed_time": "3h 44m 58s", "remaining_time": "28s"} |
|
|
{"eval_loss": 0.16983823, "eval_token_acc": 0.92398867, "eval_runtime": 1324.7962, "eval_samples_per_second": 60.675, "eval_steps_per_second": 3.792, "epoch": 0.99961034, "global_step/max_steps": "481/481", "percentage": "100.00%", "elapsed_time": "4h 7m 20s", "remaining_time": "0s"} |
|
|
{"train_runtime": 15102.7925, "train_samples_per_second": 8.156, "train_steps_per_second": 0.032, "total_flos": 5.1002858766178714e+17, "train_loss": 0.25257182, "epoch": 0.99961034, "global_step/max_steps": "481/481", "percentage": "100.00%", "elapsed_time": "4h 11m 43s", "remaining_time": "0s"} |
|
|
{"model_parameter_info": "Qwen2_5_VLForConditionalGeneration: 3754.6230M Params (3754.6230M Trainable [100.0000%]), 0.0024M Buffers.", "last_model_checkpoint": "/data4/tongshuo/Ordinal_Regression/medical_model/Qwen/Qwen-VL-master/ckp/checkpoints_IO_3B/v0-20250504-043938/checkpoint-481", "best_model_checkpoint": "/data4/tongshuo/Ordinal_Regression/medical_model/Qwen/Qwen-VL-master/ckp/checkpoints_IO_3B/v0-20250504-043938/checkpoint-400", "best_metric": 0.16972463, "global_step": 481, "log_history": [{"loss": 2.024962902069092, "token_acc": 0.5925925925925926, "grad_norm": 49.5, "learning_rate": 2.0000000000000002e-07, "memory(GiB)": 30.5, "train_speed(iter/s)": 0.096714, "epoch": 0.002078191972983504, "step": 1}, {"loss": 1.9107239246368408, "token_acc": 0.6030968247745983, "grad_norm": 48.25, "learning_rate": 1.0000000000000002e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.104169, "epoch": 0.010390959864917521, "step": 5}, {"loss": 1.8856124877929688, "token_acc": 0.6060725241878392, "grad_norm": 46.0, "learning_rate": 2.0000000000000003e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.10431, "epoch": 0.020781919729835042, "step": 10}, {"loss": 1.6671241760253905, "token_acc": 0.6276759587247805, "grad_norm": 41.25, "learning_rate": 3e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.104357, "epoch": 0.031172879594752564, "step": 15}, {"loss": 1.4283610343933106, "token_acc": 0.644587791063562, "grad_norm": 35.75, "learning_rate": 4.000000000000001e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.104096, "epoch": 0.041563839459670085, "step": 20}, {"loss": 0.8581354141235351, "token_acc": 0.7361944076641483, "grad_norm": 24.125, "learning_rate": 5e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103919, "epoch": 0.05195479932458761, "step": 25}, {"loss": 0.3941912889480591, "token_acc": 0.8509957048028114, "grad_norm": 10.4375, "learning_rate": 4.998516877114918e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103913, "epoch": 0.06234575918950513, "step": 30}, {"loss": 0.24166102409362794, "token_acc": 0.9107004193197702, "grad_norm": 3.6875, "learning_rate": 4.994069268182461e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103983, "epoch": 0.07273671905442265, "step": 35}, {"loss": 0.2008065700531006, "token_acc": 0.9205524090309566, "grad_norm": 1.828125, "learning_rate": 4.986662450283107e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103895, "epoch": 0.08312767891934017, "step": 40}, {"loss": 0.19015437364578247, "token_acc": 0.9180647678501755, "grad_norm": 2.265625, "learning_rate": 4.976305211593758e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103752, "epoch": 0.0935186387842577, "step": 45}, {"loss": 0.19012608528137206, "token_acc": 0.9183419043166589, "grad_norm": 1.9921875, "learning_rate": 4.963009840960598e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103736, "epoch": 0.10390959864917522, "step": 50}, {"loss": 0.1876815915107727, "token_acc": 0.9174203696421549, "grad_norm": 1.9921875, "learning_rate": 4.9467921133183864e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103693, "epoch": 0.11430055851409274, "step": 55}, {"loss": 0.18475005626678467, "token_acc": 0.9194869417400711, "grad_norm": 1.640625, "learning_rate": 4.927671270973533e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103749, "epoch": 0.12469151837901025, "step": 60}, {"loss": 0.18700193166732787, "token_acc": 0.9183423809159722, "grad_norm": 2.03125, "learning_rate": 4.905670000773126e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103789, "epoch": 0.13508247824392777, "step": 65}, {"loss": 0.1833273410797119, "token_acc": 0.9198803956145059, "grad_norm": 1.8984375, "learning_rate": 4.880814407187037e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.103777, "epoch": 0.1454734381088453, "step": 70}, {"loss": 0.17667144536972046, "token_acc": 0.9221237549996079, "grad_norm": 1.8046875, "learning_rate": 4.8531339813349995e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.101758, "epoch": 0.15586439797376284, "step": 75}, {"loss": 0.17785824537277223, "token_acc": 0.9208750290900628, "grad_norm": 1.8203125, "learning_rate": 4.822661565995454e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.096873, "epoch": 0.16625535783868034, "step": 80}, {"loss": 0.18061867952346802, "token_acc": 0.9204729886389984, "grad_norm": 1.5390625, "learning_rate": 4.789433316637644e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.092995, "epoch": 0.17664631770359787, "step": 85}, {"loss": 0.17701296806335448, "token_acc": 0.9225134926754048, "grad_norm": 1.8828125, "learning_rate": 4.753488658523212e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.089657, "epoch": 0.1870372775685154, "step": 90}, {"loss": 0.1849340558052063, "token_acc": 0.9169326586220319, "grad_norm": 3.21875, "learning_rate": 4.714870239928195e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.086893, "epoch": 0.1974282374334329, "step": 95}, {"loss": 0.1811493992805481, "token_acc": 0.9194093158432174, "grad_norm": 3.3125, "learning_rate": 4.673623881540917e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.084511, "epoch": 0.20781919729835044, "step": 100}, {"eval_loss": 0.17230889201164246, "eval_token_acc": 0.9252880661892655, "eval_runtime": 1320.3184, "eval_samples_per_second": 60.881, "eval_steps_per_second": 3.805, "epoch": 0.20781919729835044, "step": 100}, {"loss": 0.1755787253379822, "token_acc": 0.925254316998113, "grad_norm": 1.9375, "learning_rate": 4.629798522095818e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03748, "epoch": 0.21821015716326797, "step": 105}, {"loss": 0.16902718544006348, "token_acc": 0.9260815437286025, "grad_norm": 1.6484375, "learning_rate": 4.583446160307729e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.0381, "epoch": 0.22860111702818547, "step": 110}, {"loss": 0.17990721464157106, "token_acc": 0.9196615695102073, "grad_norm": 3.0, "learning_rate": 4.534621793175488e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038616, "epoch": 0.238992076893103, "step": 115}, {"loss": 0.17740323543548583, "token_acc": 0.9212245219170536, "grad_norm": 1.984375, "learning_rate": 4.4833833507280884e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039106, "epoch": 0.2493830367580205, "step": 120}, {"loss": 0.17831547260284425, "token_acc": 0.9183422870299309, "grad_norm": 2.03125, "learning_rate": 4.429791627290802e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039569, "epoch": 0.25977399662293804, "step": 125}, {"loss": 0.1763285517692566, "token_acc": 0.9201423817998917, "grad_norm": 1.5546875, "learning_rate": 4.373910209352816e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.040384, "epoch": 0.27016495648785555, "step": 130}, {"loss": 0.17515565156936647, "token_acc": 0.9202636948673678, "grad_norm": 1.625, "learning_rate": 4.315805400121963e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.041134, "epoch": 0.2805559163527731, "step": 135}, {"loss": 0.18066967725753785, "token_acc": 0.9194688203774171, "grad_norm": 2.09375, "learning_rate": 4.255546140856087e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.041899, "epoch": 0.2909468762176906, "step": 140}, {"loss": 0.18014354705810548, "token_acc": 0.9178993503952414, "grad_norm": 1.140625, "learning_rate": 4.1932039290643534e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.042363, "epoch": 0.3013378360826081, "step": 145}, {"loss": 0.17886133193969728, "token_acc": 0.9210793106137409, "grad_norm": 1.3515625, "learning_rate": 4.128852733675572e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.042968, "epoch": 0.31172879594752567, "step": 150}, {"loss": 0.18089871406555175, "token_acc": 0.9191895486192599, "grad_norm": 1.7734375, "learning_rate": 4.062568907274197e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.043361, "epoch": 0.3221197558124432, "step": 155}, {"loss": 0.17732292413711548, "token_acc": 0.9188111067657411, "grad_norm": 1.7421875, "learning_rate": 3.994431095508102e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.043692, "epoch": 0.3325107156773607, "step": 160}, {"loss": 0.1741779327392578, "token_acc": 0.9216403392732083, "grad_norm": 1.78125, "learning_rate": 3.9245201437756655e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.043984, "epoch": 0.34290167554227824, "step": 165}, {"loss": 0.17619532346725464, "token_acc": 0.9227966824277187, "grad_norm": 1.875, "learning_rate": 3.852919001302833e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.044279, "epoch": 0.35329263540719574, "step": 170}, {"loss": 0.17437649965286256, "token_acc": 0.9230951628201075, "grad_norm": 3.71875, "learning_rate": 3.779712622724003e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.044551, "epoch": 0.36368359527211325, "step": 175}, {"loss": 0.17724401950836183, "token_acc": 0.9202978056426332, "grad_norm": 1.84375, "learning_rate": 3.704987867283499e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.044808, "epoch": 0.3740745551370308, "step": 180}, {"loss": 0.18116596937179566, "token_acc": 0.9170886075949367, "grad_norm": 2.1875, "learning_rate": 3.628833395777224e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.045056, "epoch": 0.3844655150019483, "step": 185}, {"loss": 0.17593677043914796, "token_acc": 0.9202586206896551, "grad_norm": 2.046875, "learning_rate": 3.551339565356769e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.045325, "epoch": 0.3948564748668658, "step": 190}, {"loss": 0.1823140025138855, "token_acc": 0.917639683410391, "grad_norm": 2.0, "learning_rate": 3.4725983223208155e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.04554, "epoch": 0.4052474347317834, "step": 195}, {"loss": 0.18079570531845093, "token_acc": 0.9180621543733936, "grad_norm": 1.7109375, "learning_rate": 3.392703093021e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.045874, "epoch": 0.4156383945967009, "step": 200}, {"eval_loss": 0.17064262926578522, "eval_token_acc": 0.9238987477359346, "eval_runtime": 1310.4171, "eval_samples_per_second": 61.341, "eval_steps_per_second": 3.834, "epoch": 0.4156383945967009, "step": 200}, {"loss": 0.181607449054718, "token_acc": 0.9238299582974754, "grad_norm": 1.7890625, "learning_rate": 3.3117486730117092e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034354, "epoch": 0.4260293544616184, "step": 205}, {"loss": 0.17884130477905275, "token_acc": 0.9180160352145889, "grad_norm": 2.0625, "learning_rate": 3.229831114575315e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034671, "epoch": 0.43642031432653594, "step": 210}, {"loss": 0.174602210521698, "token_acc": 0.920204921214003, "grad_norm": 1.484375, "learning_rate": 3.147047612756302e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034976, "epoch": 0.44681127419145344, "step": 215}, {"loss": 0.1783522129058838, "token_acc": 0.9198404629702042, "grad_norm": 1.234375, "learning_rate": 3.063496390039516e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035377, "epoch": 0.45720223405637095, "step": 220}, {"loss": 0.179638671875, "token_acc": 0.9207520783155932, "grad_norm": 1.78125, "learning_rate": 2.9792765798093466e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035838, "epoch": 0.46759319392128845, "step": 225}, {"loss": 0.18421536684036255, "token_acc": 0.916609076249712, "grad_norm": 1.921875, "learning_rate": 2.8944881087281375e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036335, "epoch": 0.477984153786206, "step": 230}, {"loss": 0.17683053016662598, "token_acc": 0.9213091922005571, "grad_norm": 1.7890625, "learning_rate": 2.80923157817337e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036825, "epoch": 0.4883751136511235, "step": 235}, {"loss": 0.17713736295700072, "token_acc": 0.92035536159601, "grad_norm": 1.84375, "learning_rate": 2.723608144874298e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037143, "epoch": 0.498766073516041, "step": 240}, {"loss": 0.17735506296157838, "token_acc": 0.9191973969631236, "grad_norm": 1.5234375, "learning_rate": 2.637719400889664e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037405, "epoch": 0.5091570333809585, "step": 245}, {"loss": 0.18297756910324098, "token_acc": 0.9193918758413177, "grad_norm": 1.3828125, "learning_rate": 2.5516672530688864e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037652, "epoch": 0.5195479932458761, "step": 250}, {"loss": 0.17588781118392943, "token_acc": 0.9216273752266814, "grad_norm": 2.203125, "learning_rate": 2.4655538021397592e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037896, "epoch": 0.5299389531107936, "step": 255}, {"loss": 0.18080203533172606, "token_acc": 0.921203216826477, "grad_norm": 1.6796875, "learning_rate": 2.3794812215661134e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038125, "epoch": 0.5403299129757111, "step": 260}, {"loss": 0.18104695081710814, "token_acc": 0.9180170831005029, "grad_norm": 2.6875, "learning_rate": 2.2935516363191695e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038349, "epoch": 0.5507208728406287, "step": 265}, {"loss": 0.1768229365348816, "token_acc": 0.9203719621786356, "grad_norm": 1.7578125, "learning_rate": 2.2078670017064366e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038567, "epoch": 0.5611118327055462, "step": 270}, {"loss": 0.17726024389266967, "token_acc": 0.9191293295201779, "grad_norm": 1.34375, "learning_rate": 2.1225289824019077e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.038795, "epoch": 0.5715027925704637, "step": 275}, {"loss": 0.18150538206100464, "token_acc": 0.9174561267277528, "grad_norm": 1.71875, "learning_rate": 2.037638831821104e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03901, "epoch": 0.5818937524353812, "step": 280}, {"loss": 0.1750028371810913, "token_acc": 0.9228211275230067, "grad_norm": 1.546875, "learning_rate": 1.953297271984061e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039295, "epoch": 0.5922847123002988, "step": 285}, {"loss": 0.17536247968673707, "token_acc": 0.9208355815925383, "grad_norm": 1.234375, "learning_rate": 1.8696043740088236e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039587, "epoch": 0.6026756721652162, "step": 290}, {"loss": 0.1767476439476013, "token_acc": 0.9220417633410672, "grad_norm": 1.5390625, "learning_rate": 1.7866594393772375e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.039803, "epoch": 0.6130666320301338, "step": 295}, {"loss": 0.17351619005203248, "token_acc": 0.9228345246330029, "grad_norm": 1.9140625, "learning_rate": 1.7045608821139045e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.040008, "epoch": 0.6234575918950513, "step": 300}, {"eval_loss": 0.16996929049491882, "eval_token_acc": 0.9252277899263542, "eval_runtime": 1338.6589, "eval_samples_per_second": 60.047, "eval_steps_per_second": 3.753, "epoch": 0.6234575918950513, "step": 300}, {"loss": 0.1729782223701477, "token_acc": 0.9252146572948351, "grad_norm": 1.7109375, "learning_rate": 1.6234061120181144e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033422, "epoch": 0.6338485517599688, "step": 305}, {"loss": 0.1758588194847107, "token_acc": 0.9202410211686355, "grad_norm": 2.296875, "learning_rate": 1.5432914190872757e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033685, "epoch": 0.6442395116248864, "step": 310}, {"loss": 0.17637253999710084, "token_acc": 0.9202949713658115, "grad_norm": 2.03125, "learning_rate": 1.464311859269003e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034017, "epoch": 0.6546304714898039, "step": 315}, {"loss": 0.1803775191307068, "token_acc": 0.9197953561589925, "grad_norm": 2.296875, "learning_rate": 1.3865611416773921e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034317, "epoch": 0.6650214313547214, "step": 320}, {"loss": 0.17441234588623047, "token_acc": 0.9233814625058221, "grad_norm": 2.046875, "learning_rate": 1.3101315174073162e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034642, "epoch": 0.6754123912196389, "step": 325}, {"loss": 0.17859526872634887, "token_acc": 0.9203665665981988, "grad_norm": 1.640625, "learning_rate": 1.235113670078658e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034906, "epoch": 0.6858033510845565, "step": 330}, {"loss": 0.17650686502456664, "token_acc": 0.9200734956361966, "grad_norm": 1.375, "learning_rate": 1.161596608240349e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035109, "epoch": 0.6961943109494739, "step": 335}, {"loss": 0.18125290870666505, "token_acc": 0.9172004012655297, "grad_norm": 1.7734375, "learning_rate": 1.0896675597618725e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035306, "epoch": 0.7065852708143915, "step": 340}, {"loss": 0.1810152769088745, "token_acc": 0.9158885657431501, "grad_norm": 1.046875, "learning_rate": 1.0194118683375502e-06, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035495, "epoch": 0.716976230679309, "step": 345}, {"loss": 0.17332950830459595, "token_acc": 0.9222935923546921, "grad_norm": 2.203125, "learning_rate": 9.509128922263886e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035686, "epoch": 0.7273671905442265, "step": 350}, {"loss": 0.17291887998580932, "token_acc": 0.9233901951744276, "grad_norm": 1.28125, "learning_rate": 8.842519053476476e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035869, "epoch": 0.737758150409144, "step": 355}, {"loss": 0.17527254819869995, "token_acc": 0.9194406927782632, "grad_norm": 1.5859375, "learning_rate": 8.195080008494744e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03605, "epoch": 0.7481491102740616, "step": 360}, {"loss": 0.1758326292037964, "token_acc": 0.9222837796244998, "grad_norm": 1.8984375, "learning_rate": 7.567579972650116e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036228, "epoch": 0.7585400701389791, "step": 365}, {"loss": 0.17296208143234254, "token_acc": 0.9199661069172701, "grad_norm": 1.4453125, "learning_rate": 6.960763473673451e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036459, "epoch": 0.7689310300038966, "step": 370}, {"loss": 0.17531843185424806, "token_acc": 0.921526620645878, "grad_norm": 2.15625, "learning_rate": 6.375350498314075e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036681, "epoch": 0.7793219898688142, "step": 375}, {"loss": 0.17355836629867555, "token_acc": 0.9211911683821796, "grad_norm": 1.453125, "learning_rate": 5.812035638076785e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.036859, "epoch": 0.7897129497337316, "step": 380}, {"loss": 0.1834742784500122, "token_acc": 0.9166990442147797, "grad_norm": 2.03125, "learning_rate": 5.271487265090163e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037022, "epoch": 0.8001039095986492, "step": 385}, {"loss": 0.17866382598876954, "token_acc": 0.9190055507778907, "grad_norm": 1.78125, "learning_rate": 4.754346739084173e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037179, "epoch": 0.8104948694635667, "step": 390}, {"loss": 0.16837071180343627, "token_acc": 0.925252047889099, "grad_norm": 1.7265625, "learning_rate": 4.2612276464179673e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037337, "epoch": 0.8208858293284842, "step": 395}, {"loss": 0.17911137342453004, "token_acc": 0.920388196307401, "grad_norm": 1.625, "learning_rate": 3.7927150720606596e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.037494, "epoch": 0.8312767891934018, "step": 400}, {"eval_loss": 0.16972462832927704, "eval_token_acc": 0.9239007240068497, "eval_runtime": 1325.8316, "eval_samples_per_second": 60.628, "eval_steps_per_second": 3.789, "epoch": 0.8312767891934018, "step": 400}, {"loss": 0.17124630212783815, "token_acc": 0.9238681922697054, "grad_norm": 1.546875, "learning_rate": 3.3493649053890325e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.032973, "epoch": 0.8416677490583193, "step": 405}, {"loss": 0.17123721837997435, "token_acc": 0.9216517506672947, "grad_norm": 2.25, "learning_rate": 2.931703180625736e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033214, "epoch": 0.8520587089232368, "step": 410}, {"loss": 0.17746515274047853, "token_acc": 0.9180823830579474, "grad_norm": 1.5859375, "learning_rate": 2.5402254527005286e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033458, "epoch": 0.8624496687881543, "step": 415}, {"loss": 0.1777236580848694, "token_acc": 0.9190801349972529, "grad_norm": 1.3984375, "learning_rate": 2.1753962092752545e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033705, "epoch": 0.8728406286530719, "step": 420}, {"loss": 0.17560485601425171, "token_acc": 0.9201813013751249, "grad_norm": 1.53125, "learning_rate": 1.837648319629956e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.033868, "epoch": 0.8832315885179893, "step": 425}, {"loss": 0.17915327548980714, "token_acc": 0.9182488623882002, "grad_norm": 1.7578125, "learning_rate": 1.5273825210642608e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034029, "epoch": 0.8936225483829069, "step": 430}, {"loss": 0.17606736421585084, "token_acc": 0.9228688461241209, "grad_norm": 1.3671875, "learning_rate": 1.2449669434232202e-07, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034183, "epoch": 0.9040135082478243, "step": 435}, {"loss": 0.17585525512695313, "token_acc": 0.9183866393571766, "grad_norm": 1.546875, "learning_rate": 9.907366723118678e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034333, "epoch": 0.9144044681127419, "step": 440}, {"loss": 0.16930280923843383, "token_acc": 0.9238362821778419, "grad_norm": 1.5, "learning_rate": 7.649933515167407e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.034485, "epoch": 0.9247954279776595, "step": 445}, {"loss": 0.1705850124359131, "token_acc": 0.925625, "grad_norm": 2.078125, "learning_rate": 5.6800482510601937e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03464, "epoch": 0.9351863878425769, "step": 450}, {"loss": 0.1817856550216675, "token_acc": 0.9208294062205467, "grad_norm": 1.5703125, "learning_rate": 4.000048196330014e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03485, "epoch": 0.9455773477074945, "step": 455}, {"loss": 0.1783198595046997, "token_acc": 0.9211605415860735, "grad_norm": 1.234375, "learning_rate": 2.611926668199316e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035001, "epoch": 0.955968307572412, "step": 460}, {"loss": 0.17677730321884155, "token_acc": 0.9180868609125893, "grad_norm": 1.59375, "learning_rate": 1.517330670512629e-08, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03514, "epoch": 0.9663592674373295, "step": 465}, {"loss": 0.17313094139099122, "token_acc": 0.9224636548382054, "grad_norm": 1.984375, "learning_rate": 7.175589395692351e-09, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03528, "epoch": 0.976750227302247, "step": 470}, {"loss": 0.1789810061454773, "token_acc": 0.918305714728202, "grad_norm": 2.0625, "learning_rate": 2.1356040317474512e-09, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.035421, "epoch": 0.9871411871671646, "step": 475}, {"loss": 0.17718052864074707, "token_acc": 0.9196187450357427, "grad_norm": 1.296875, "learning_rate": 5.933054739837296e-11, "memory(GiB)": 36.99, "train_speed(iter/s)": 0.03556, "epoch": 0.997532147032082, "step": 480}, {"eval_loss": 0.16983823478221893, "eval_token_acc": 0.9239886680625727, "eval_runtime": 1324.7962, "eval_samples_per_second": 60.675, "eval_steps_per_second": 3.792, "epoch": 0.9996103390050656, "step": 481}, {"train_runtime": 15102.7925, "train_samples_per_second": 8.156, "train_steps_per_second": 0.032, "total_flos": 5.1002858766178714e+17, "train_loss": 0.2525718175571822, "epoch": 0.9996103390050656, "step": 481}], "memory": 36.98828125} |
|
|
|