| [ | |
| { | |
| "current_steps": 63, | |
| "loss": 2.8305, | |
| "learning_rate": 0.0005, | |
| "epoch": 0.02549800796812749 | |
| }, | |
| { | |
| "current_steps": 127, | |
| "loss": 2.7784, | |
| "learning_rate": 0.001, | |
| "epoch": 0.05099601593625498 | |
| }, | |
| { | |
| "current_steps": 191, | |
| "loss": 2.6623, | |
| "learning_rate": 0.0009974226804123712, | |
| "epoch": 0.07649402390438247 | |
| }, | |
| { | |
| "current_steps": 255, | |
| "loss": 2.6442, | |
| "learning_rate": 0.0009948453608247423, | |
| "epoch": 0.10199203187250996 | |
| }, | |
| { | |
| "current_steps": 319, | |
| "loss": 2.822, | |
| "learning_rate": 0.0009922680412371135, | |
| "epoch": 0.12749003984063745 | |
| }, | |
| { | |
| "current_steps": 383, | |
| "loss": 2.6869, | |
| "learning_rate": 0.0009896907216494846, | |
| "epoch": 0.15298804780876493 | |
| }, | |
| { | |
| "current_steps": 447, | |
| "loss": 2.5865, | |
| "learning_rate": 0.0009871134020618558, | |
| "epoch": 0.17848605577689244 | |
| }, | |
| { | |
| "current_steps": 511, | |
| "loss": 2.6222, | |
| "learning_rate": 0.000984536082474227, | |
| "epoch": 0.20398406374501993 | |
| }, | |
| { | |
| "current_steps": 575, | |
| "loss": 2.5421, | |
| "learning_rate": 0.0009819587628865979, | |
| "epoch": 0.2294820717131474 | |
| }, | |
| { | |
| "current_steps": 639, | |
| "loss": 2.5766, | |
| "learning_rate": 0.000979381443298969, | |
| "epoch": 0.2549800796812749 | |
| }, | |
| { | |
| "current_steps": 703, | |
| "loss": 2.5266, | |
| "learning_rate": 0.0009768041237113402, | |
| "epoch": 0.2804780876494024 | |
| }, | |
| { | |
| "current_steps": 767, | |
| "loss": 2.5239, | |
| "learning_rate": 0.0009742268041237113, | |
| "epoch": 0.30597609561752986 | |
| }, | |
| { | |
| "current_steps": 831, | |
| "loss": 2.4799, | |
| "learning_rate": 0.0009716494845360825, | |
| "epoch": 0.3314741035856574 | |
| }, | |
| { | |
| "current_steps": 895, | |
| "loss": 2.5332, | |
| "learning_rate": 0.0009690721649484536, | |
| "epoch": 0.3569721115537849 | |
| }, | |
| { | |
| "current_steps": 959, | |
| "loss": 2.4283, | |
| "learning_rate": 0.0009664948453608248, | |
| "epoch": 0.38247011952191234 | |
| }, | |
| { | |
| "current_steps": 1023, | |
| "loss": 2.4595, | |
| "learning_rate": 0.0009639175257731959, | |
| "epoch": 0.40796812749003986 | |
| }, | |
| { | |
| "current_steps": 1087, | |
| "loss": 2.4821, | |
| "learning_rate": 0.0009613402061855671, | |
| "epoch": 0.4334661354581673 | |
| }, | |
| { | |
| "current_steps": 1151, | |
| "loss": 2.4473, | |
| "learning_rate": 0.0009587628865979382, | |
| "epoch": 0.4589641434262948 | |
| }, | |
| { | |
| "current_steps": 1215, | |
| "loss": 2.4614, | |
| "learning_rate": 0.0009561855670103094, | |
| "epoch": 0.48446215139442234 | |
| }, | |
| { | |
| "current_steps": 1279, | |
| "loss": 2.4037, | |
| "learning_rate": 0.0009536082474226805, | |
| "epoch": 0.5099601593625498 | |
| }, | |
| { | |
| "current_steps": 1343, | |
| "loss": 2.4243, | |
| "learning_rate": 0.0009510309278350515, | |
| "epoch": 0.5354581673306773 | |
| }, | |
| { | |
| "current_steps": 1407, | |
| "loss": 2.3534, | |
| "learning_rate": 0.0009484536082474226, | |
| "epoch": 0.5609561752988048 | |
| }, | |
| { | |
| "current_steps": 1471, | |
| "loss": 2.406, | |
| "learning_rate": 0.0009458762886597938, | |
| "epoch": 0.5864541832669322 | |
| }, | |
| { | |
| "current_steps": 1535, | |
| "loss": 2.483, | |
| "learning_rate": 0.0009432989690721649, | |
| "epoch": 0.6119521912350597 | |
| }, | |
| { | |
| "current_steps": 1599, | |
| "loss": 2.4313, | |
| "learning_rate": 0.0009407216494845361, | |
| "epoch": 0.6374501992031872 | |
| }, | |
| { | |
| "current_steps": 1663, | |
| "loss": 2.4101, | |
| "learning_rate": 0.0009381443298969072, | |
| "epoch": 0.6629482071713148 | |
| }, | |
| { | |
| "current_steps": 1727, | |
| "loss": 2.3847, | |
| "learning_rate": 0.0009355670103092784, | |
| "epoch": 0.6884462151394423 | |
| }, | |
| { | |
| "current_steps": 1791, | |
| "loss": 2.3194, | |
| "learning_rate": 0.0009329896907216495, | |
| "epoch": 0.7139442231075698 | |
| }, | |
| { | |
| "current_steps": 1855, | |
| "loss": 2.3921, | |
| "learning_rate": 0.0009304123711340207, | |
| "epoch": 0.7394422310756972 | |
| }, | |
| { | |
| "current_steps": 1919, | |
| "loss": 2.4158, | |
| "learning_rate": 0.0009278350515463918, | |
| "epoch": 0.7649402390438247 | |
| }, | |
| { | |
| "current_steps": 1983, | |
| "loss": 2.3134, | |
| "learning_rate": 0.000925257731958763, | |
| "epoch": 0.7904382470119522 | |
| }, | |
| { | |
| "current_steps": 2047, | |
| "loss": 2.3514, | |
| "learning_rate": 0.0009226804123711341, | |
| "epoch": 0.8159362549800797 | |
| }, | |
| { | |
| "current_steps": 2111, | |
| "loss": 2.3011, | |
| "learning_rate": 0.0009201030927835051, | |
| "epoch": 0.8414342629482072 | |
| }, | |
| { | |
| "current_steps": 2175, | |
| "loss": 2.4016, | |
| "learning_rate": 0.0009175257731958762, | |
| "epoch": 0.8669322709163346 | |
| }, | |
| { | |
| "current_steps": 2239, | |
| "loss": 2.3209, | |
| "learning_rate": 0.0009149484536082474, | |
| "epoch": 0.8924302788844621 | |
| }, | |
| { | |
| "current_steps": 2303, | |
| "loss": 2.3672, | |
| "learning_rate": 0.0009123711340206185, | |
| "epoch": 0.9179282868525896 | |
| }, | |
| { | |
| "current_steps": 2367, | |
| "loss": 2.3597, | |
| "learning_rate": 0.0009097938144329897, | |
| "epoch": 0.9434262948207172 | |
| }, | |
| { | |
| "current_steps": 2431, | |
| "loss": 2.3619, | |
| "learning_rate": 0.0009072164948453608, | |
| "epoch": 0.9689243027888447 | |
| }, | |
| { | |
| "current_steps": 2495, | |
| "loss": 2.3521, | |
| "learning_rate": 0.000904639175257732, | |
| "epoch": 0.9944223107569721 | |
| }, | |
| { | |
| "current_steps": 2545, | |
| "loss": 2.2467, | |
| "learning_rate": 0.0009020618556701031, | |
| "epoch": 1.0199203187250996 | |
| }, | |
| { | |
| "current_steps": 2609, | |
| "loss": 2.2082, | |
| "learning_rate": 0.0008994845360824743, | |
| "epoch": 1.045418326693227 | |
| }, | |
| { | |
| "current_steps": 2673, | |
| "loss": 2.2161, | |
| "learning_rate": 0.0008969072164948454, | |
| "epoch": 1.0709163346613546 | |
| }, | |
| { | |
| "current_steps": 2737, | |
| "loss": 2.1961, | |
| "learning_rate": 0.0008943298969072166, | |
| "epoch": 1.0964143426294821 | |
| }, | |
| { | |
| "current_steps": 2801, | |
| "loss": 2.215, | |
| "learning_rate": 0.0008917525773195877, | |
| "epoch": 1.1219123505976096 | |
| }, | |
| { | |
| "current_steps": 2865, | |
| "loss": 2.1951, | |
| "learning_rate": 0.0008891752577319587, | |
| "epoch": 1.1474103585657371 | |
| }, | |
| { | |
| "current_steps": 2929, | |
| "loss": 2.1599, | |
| "learning_rate": 0.0008865979381443298, | |
| "epoch": 1.1729083665338647 | |
| }, | |
| { | |
| "current_steps": 2993, | |
| "loss": 2.1511, | |
| "learning_rate": 0.000884020618556701, | |
| "epoch": 1.198406374501992 | |
| }, | |
| { | |
| "current_steps": 3057, | |
| "loss": 2.0713, | |
| "learning_rate": 0.0008814432989690721, | |
| "epoch": 1.2239043824701195 | |
| }, | |
| { | |
| "current_steps": 3121, | |
| "loss": 2.1312, | |
| "learning_rate": 0.0008788659793814433, | |
| "epoch": 1.249402390438247 | |
| }, | |
| { | |
| "current_steps": 3185, | |
| "loss": 2.1442, | |
| "learning_rate": 0.0008762886597938144, | |
| "epoch": 1.2749003984063745 | |
| }, | |
| { | |
| "current_steps": 3249, | |
| "loss": 2.141, | |
| "learning_rate": 0.0008737113402061856, | |
| "epoch": 1.300398406374502 | |
| }, | |
| { | |
| "current_steps": 3313, | |
| "loss": 2.1281, | |
| "learning_rate": 0.0008711340206185567, | |
| "epoch": 1.3258964143426295 | |
| }, | |
| { | |
| "current_steps": 3377, | |
| "loss": 2.1373, | |
| "learning_rate": 0.0008685567010309279, | |
| "epoch": 1.351394422310757 | |
| }, | |
| { | |
| "current_steps": 3441, | |
| "loss": 2.1111, | |
| "learning_rate": 0.000865979381443299, | |
| "epoch": 1.3768924302788845 | |
| }, | |
| { | |
| "current_steps": 3505, | |
| "loss": 2.0973, | |
| "learning_rate": 0.0008634020618556702, | |
| "epoch": 1.402390438247012 | |
| }, | |
| { | |
| "current_steps": 3569, | |
| "loss": 2.1299, | |
| "learning_rate": 0.0008608247422680414, | |
| "epoch": 1.4278884462151393 | |
| }, | |
| { | |
| "current_steps": 3633, | |
| "loss": 2.0971, | |
| "learning_rate": 0.0008582474226804123, | |
| "epoch": 1.453386454183267 | |
| }, | |
| { | |
| "current_steps": 3697, | |
| "loss": 2.0334, | |
| "learning_rate": 0.0008556701030927834, | |
| "epoch": 1.4788844621513944 | |
| }, | |
| { | |
| "current_steps": 3761, | |
| "loss": 2.0659, | |
| "learning_rate": 0.0008530927835051546, | |
| "epoch": 1.5043824701195219 | |
| }, | |
| { | |
| "current_steps": 3825, | |
| "loss": 2.0648, | |
| "learning_rate": 0.0008505154639175257, | |
| "epoch": 1.5298804780876494 | |
| }, | |
| { | |
| "current_steps": 3889, | |
| "loss": 2.0629, | |
| "learning_rate": 0.0008479381443298969, | |
| "epoch": 1.5553784860557769 | |
| }, | |
| { | |
| "current_steps": 3953, | |
| "loss": 2.0136, | |
| "learning_rate": 0.000845360824742268, | |
| "epoch": 1.5808764940239044 | |
| }, | |
| { | |
| "current_steps": 4017, | |
| "loss": 2.0692, | |
| "learning_rate": 0.0008427835051546392, | |
| "epoch": 1.606374501992032 | |
| }, | |
| { | |
| "current_steps": 4081, | |
| "loss": 2.1114, | |
| "learning_rate": 0.0008402061855670104, | |
| "epoch": 1.6318725099601594 | |
| }, | |
| { | |
| "current_steps": 4145, | |
| "loss": 2.048, | |
| "learning_rate": 0.0008376288659793815, | |
| "epoch": 1.6573705179282867 | |
| }, | |
| { | |
| "current_steps": 4209, | |
| "loss": 2.1087, | |
| "learning_rate": 0.0008350515463917527, | |
| "epoch": 1.6828685258964144 | |
| }, | |
| { | |
| "current_steps": 4273, | |
| "loss": 2.0474, | |
| "learning_rate": 0.0008324742268041238, | |
| "epoch": 1.7083665338645417 | |
| }, | |
| { | |
| "current_steps": 4337, | |
| "loss": 2.0699, | |
| "learning_rate": 0.000829896907216495, | |
| "epoch": 1.7338645418326695 | |
| }, | |
| { | |
| "current_steps": 4401, | |
| "loss": 2.0901, | |
| "learning_rate": 0.0008273195876288659, | |
| "epoch": 1.7593625498007968 | |
| }, | |
| { | |
| "current_steps": 4465, | |
| "loss": 2.018, | |
| "learning_rate": 0.000824742268041237, | |
| "epoch": 1.7848605577689243 | |
| }, | |
| { | |
| "current_steps": 4529, | |
| "loss": 2.039, | |
| "learning_rate": 0.0008221649484536082, | |
| "epoch": 1.8103585657370518 | |
| }, | |
| { | |
| "current_steps": 4593, | |
| "loss": 2.0168, | |
| "learning_rate": 0.0008195876288659793, | |
| "epoch": 1.8358565737051793 | |
| }, | |
| { | |
| "current_steps": 4657, | |
| "loss": 1.9646, | |
| "learning_rate": 0.0008170103092783505, | |
| "epoch": 1.8613545816733068 | |
| }, | |
| { | |
| "current_steps": 4721, | |
| "loss": 1.9875, | |
| "learning_rate": 0.0008144329896907217, | |
| "epoch": 1.886852589641434 | |
| }, | |
| { | |
| "current_steps": 4785, | |
| "loss": 2.0373, | |
| "learning_rate": 0.0008118556701030928, | |
| "epoch": 1.9123505976095618 | |
| }, | |
| { | |
| "current_steps": 4849, | |
| "loss": 1.9158, | |
| "learning_rate": 0.000809278350515464, | |
| "epoch": 1.9378486055776891 | |
| }, | |
| { | |
| "current_steps": 4913, | |
| "loss": 1.9173, | |
| "learning_rate": 0.0008067010309278351, | |
| "epoch": 1.9633466135458169 | |
| }, | |
| { | |
| "current_steps": 4977, | |
| "loss": 1.9941, | |
| "learning_rate": 0.0008041237113402063, | |
| "epoch": 1.9888446215139441 | |
| }, | |
| { | |
| "current_steps": 5027, | |
| "loss": 1.8231, | |
| "learning_rate": 0.0008015463917525774, | |
| "epoch": 2.014342629482072 | |
| }, | |
| { | |
| "current_steps": 5091, | |
| "loss": 1.8018, | |
| "learning_rate": 0.0007989690721649486, | |
| "epoch": 2.039840637450199 | |
| }, | |
| { | |
| "current_steps": 5155, | |
| "loss": 1.7098, | |
| "learning_rate": 0.0007963917525773195, | |
| "epoch": 2.065338645418327 | |
| }, | |
| { | |
| "current_steps": 5219, | |
| "loss": 1.7214, | |
| "learning_rate": 0.0007938144329896907, | |
| "epoch": 2.090836653386454 | |
| }, | |
| { | |
| "current_steps": 5283, | |
| "loss": 1.8029, | |
| "learning_rate": 0.0007912371134020618, | |
| "epoch": 2.1163346613545815 | |
| }, | |
| { | |
| "current_steps": 5347, | |
| "loss": 1.71, | |
| "learning_rate": 0.000788659793814433, | |
| "epoch": 2.141832669322709 | |
| }, | |
| { | |
| "current_steps": 5411, | |
| "loss": 1.725, | |
| "learning_rate": 0.0007860824742268041, | |
| "epoch": 2.1673306772908365 | |
| }, | |
| { | |
| "current_steps": 5475, | |
| "loss": 1.729, | |
| "learning_rate": 0.0007835051546391753, | |
| "epoch": 2.1928286852589642 | |
| }, | |
| { | |
| "current_steps": 5539, | |
| "loss": 1.7139, | |
| "learning_rate": 0.0007809278350515464, | |
| "epoch": 2.2183266932270915 | |
| }, | |
| { | |
| "current_steps": 5603, | |
| "loss": 1.6588, | |
| "learning_rate": 0.0007783505154639176, | |
| "epoch": 2.2438247011952193 | |
| }, | |
| { | |
| "current_steps": 5667, | |
| "loss": 1.7179, | |
| "learning_rate": 0.0007757731958762887, | |
| "epoch": 2.2693227091633466 | |
| }, | |
| { | |
| "current_steps": 5731, | |
| "loss": 1.7024, | |
| "learning_rate": 0.0007731958762886599, | |
| "epoch": 2.2948207171314743 | |
| }, | |
| { | |
| "current_steps": 5795, | |
| "loss": 1.7205, | |
| "learning_rate": 0.000770618556701031, | |
| "epoch": 2.3203187250996016 | |
| }, | |
| { | |
| "current_steps": 5859, | |
| "loss": 1.7094, | |
| "learning_rate": 0.0007680412371134022, | |
| "epoch": 2.3458167330677293 | |
| }, | |
| { | |
| "current_steps": 5923, | |
| "loss": 1.7152, | |
| "learning_rate": 0.0007654639175257731, | |
| "epoch": 2.3713147410358566 | |
| }, | |
| { | |
| "current_steps": 5987, | |
| "loss": 1.7285, | |
| "learning_rate": 0.0007628865979381443, | |
| "epoch": 2.396812749003984 | |
| }, | |
| { | |
| "current_steps": 6051, | |
| "loss": 1.6915, | |
| "learning_rate": 0.0007603092783505154, | |
| "epoch": 2.4223107569721116 | |
| }, | |
| { | |
| "current_steps": 6115, | |
| "loss": 1.6536, | |
| "learning_rate": 0.0007577319587628866, | |
| "epoch": 2.447808764940239 | |
| }, | |
| { | |
| "current_steps": 6179, | |
| "loss": 1.6924, | |
| "learning_rate": 0.0007551546391752577, | |
| "epoch": 2.4733067729083666 | |
| }, | |
| { | |
| "current_steps": 6243, | |
| "loss": 1.6518, | |
| "learning_rate": 0.0007525773195876289, | |
| "epoch": 2.498804780876494 | |
| }, | |
| { | |
| "current_steps": 6307, | |
| "loss": 1.6765, | |
| "learning_rate": 0.00075, | |
| "epoch": 2.5243027888446217 | |
| }, | |
| { | |
| "current_steps": 6371, | |
| "loss": 1.6662, | |
| "learning_rate": 0.0007474226804123712, | |
| "epoch": 2.549800796812749 | |
| }, | |
| { | |
| "current_steps": 6435, | |
| "loss": 1.6179, | |
| "learning_rate": 0.0007448453608247423, | |
| "epoch": 2.5752988047808767 | |
| }, | |
| { | |
| "current_steps": 6499, | |
| "loss": 1.6485, | |
| "learning_rate": 0.0007422680412371135, | |
| "epoch": 2.600796812749004 | |
| }, | |
| { | |
| "current_steps": 6563, | |
| "loss": 1.7025, | |
| "learning_rate": 0.0007396907216494846, | |
| "epoch": 2.6262948207171313 | |
| }, | |
| { | |
| "current_steps": 6627, | |
| "loss": 1.6723, | |
| "learning_rate": 0.0007371134020618558, | |
| "epoch": 2.651792828685259 | |
| }, | |
| { | |
| "current_steps": 6691, | |
| "loss": 1.7608, | |
| "learning_rate": 0.0007345360824742269, | |
| "epoch": 2.6772908366533863 | |
| }, | |
| { | |
| "current_steps": 6755, | |
| "loss": 1.6388, | |
| "learning_rate": 0.0007319587628865979, | |
| "epoch": 2.702788844621514 | |
| }, | |
| { | |
| "current_steps": 6819, | |
| "loss": 1.6381, | |
| "learning_rate": 0.000729381443298969, | |
| "epoch": 2.7282868525896413 | |
| }, | |
| { | |
| "current_steps": 6883, | |
| "loss": 1.6857, | |
| "learning_rate": 0.0007268041237113402, | |
| "epoch": 2.753784860557769 | |
| }, | |
| { | |
| "current_steps": 6947, | |
| "loss": 1.6556, | |
| "learning_rate": 0.0007242268041237113, | |
| "epoch": 2.7792828685258963 | |
| }, | |
| { | |
| "current_steps": 7011, | |
| "loss": 1.6446, | |
| "learning_rate": 0.0007216494845360825, | |
| "epoch": 2.804780876494024 | |
| }, | |
| { | |
| "current_steps": 7075, | |
| "loss": 1.593, | |
| "learning_rate": 0.0007190721649484536, | |
| "epoch": 2.8302788844621514 | |
| }, | |
| { | |
| "current_steps": 7139, | |
| "loss": 1.6836, | |
| "learning_rate": 0.0007164948453608248, | |
| "epoch": 2.8557768924302787 | |
| }, | |
| { | |
| "current_steps": 7203, | |
| "loss": 1.5774, | |
| "learning_rate": 0.0007139175257731959, | |
| "epoch": 2.8812749003984064 | |
| }, | |
| { | |
| "current_steps": 7267, | |
| "loss": 1.6251, | |
| "learning_rate": 0.0007113402061855671, | |
| "epoch": 2.906772908366534 | |
| }, | |
| { | |
| "current_steps": 7331, | |
| "loss": 1.6695, | |
| "learning_rate": 0.0007087628865979382, | |
| "epoch": 2.9322709163346614 | |
| }, | |
| { | |
| "current_steps": 7395, | |
| "loss": 1.6645, | |
| "learning_rate": 0.0007061855670103094, | |
| "epoch": 2.9577689243027887 | |
| }, | |
| { | |
| "current_steps": 7459, | |
| "loss": 1.6589, | |
| "learning_rate": 0.0007036082474226805, | |
| "epoch": 2.9832669322709164 | |
| }, | |
| { | |
| "current_steps": 7509, | |
| "loss": 1.4876, | |
| "learning_rate": 0.0007010309278350515, | |
| "epoch": 3.0087649402390437 | |
| }, | |
| { | |
| "current_steps": 7573, | |
| "loss": 1.3509, | |
| "learning_rate": 0.0006984536082474226, | |
| "epoch": 3.0342629482071715 | |
| }, | |
| { | |
| "current_steps": 7637, | |
| "loss": 1.3058, | |
| "learning_rate": 0.0006958762886597938, | |
| "epoch": 3.0597609561752988 | |
| }, | |
| { | |
| "current_steps": 7701, | |
| "loss": 1.4409, | |
| "learning_rate": 0.0006932989690721649, | |
| "epoch": 3.0852589641434265 | |
| }, | |
| { | |
| "current_steps": 7765, | |
| "loss": 1.3829, | |
| "learning_rate": 0.0006907216494845361, | |
| "epoch": 3.1107569721115538 | |
| }, | |
| { | |
| "current_steps": 7829, | |
| "loss": 1.3513, | |
| "learning_rate": 0.0006881443298969072, | |
| "epoch": 3.1362549800796815 | |
| }, | |
| { | |
| "current_steps": 7893, | |
| "loss": 1.3223, | |
| "learning_rate": 0.0006855670103092784, | |
| "epoch": 3.161752988047809 | |
| }, | |
| { | |
| "current_steps": 7957, | |
| "loss": 1.2705, | |
| "learning_rate": 0.0006829896907216495, | |
| "epoch": 3.187250996015936 | |
| }, | |
| { | |
| "current_steps": 8021, | |
| "loss": 1.3133, | |
| "learning_rate": 0.0006804123711340207, | |
| "epoch": 3.212749003984064 | |
| }, | |
| { | |
| "current_steps": 8085, | |
| "loss": 1.3229, | |
| "learning_rate": 0.0006778350515463918, | |
| "epoch": 3.238247011952191 | |
| }, | |
| { | |
| "current_steps": 8149, | |
| "loss": 1.3097, | |
| "learning_rate": 0.000675257731958763, | |
| "epoch": 3.263745019920319 | |
| }, | |
| { | |
| "current_steps": 8213, | |
| "loss": 1.2961, | |
| "learning_rate": 0.0006726804123711341, | |
| "epoch": 3.289243027888446 | |
| }, | |
| { | |
| "current_steps": 8277, | |
| "loss": 1.3048, | |
| "learning_rate": 0.0006701030927835051, | |
| "epoch": 3.314741035856574 | |
| }, | |
| { | |
| "current_steps": 8341, | |
| "loss": 1.2909, | |
| "learning_rate": 0.0006675257731958762, | |
| "epoch": 3.340239043824701 | |
| }, | |
| { | |
| "current_steps": 8405, | |
| "loss": 1.3333, | |
| "learning_rate": 0.0006649484536082474, | |
| "epoch": 3.365737051792829 | |
| }, | |
| { | |
| "current_steps": 8469, | |
| "loss": 1.2552, | |
| "learning_rate": 0.0006623711340206185, | |
| "epoch": 3.391235059760956 | |
| }, | |
| { | |
| "current_steps": 8533, | |
| "loss": 1.306, | |
| "learning_rate": 0.0006597938144329897, | |
| "epoch": 3.4167330677290835 | |
| }, | |
| { | |
| "current_steps": 8597, | |
| "loss": 1.2382, | |
| "learning_rate": 0.0006572164948453608, | |
| "epoch": 3.442231075697211 | |
| }, | |
| { | |
| "current_steps": 8661, | |
| "loss": 1.2718, | |
| "learning_rate": 0.000654639175257732, | |
| "epoch": 3.4677290836653385 | |
| }, | |
| { | |
| "current_steps": 8725, | |
| "loss": 1.2348, | |
| "learning_rate": 0.0006520618556701031, | |
| "epoch": 3.4932270916334662 | |
| }, | |
| { | |
| "current_steps": 8789, | |
| "loss": 1.2724, | |
| "learning_rate": 0.0006494845360824743, | |
| "epoch": 3.5187250996015935 | |
| }, | |
| { | |
| "current_steps": 8853, | |
| "loss": 1.244, | |
| "learning_rate": 0.0006469072164948454, | |
| "epoch": 3.5442231075697213 | |
| }, | |
| { | |
| "current_steps": 8917, | |
| "loss": 1.2948, | |
| "learning_rate": 0.0006443298969072166, | |
| "epoch": 3.5697211155378485 | |
| }, | |
| { | |
| "current_steps": 8981, | |
| "loss": 1.2063, | |
| "learning_rate": 0.0006417525773195877, | |
| "epoch": 3.5952191235059763 | |
| }, | |
| { | |
| "current_steps": 9045, | |
| "loss": 1.2656, | |
| "learning_rate": 0.0006391752577319587, | |
| "epoch": 3.6207171314741036 | |
| }, | |
| { | |
| "current_steps": 9109, | |
| "loss": 1.25, | |
| "learning_rate": 0.0006365979381443298, | |
| "epoch": 3.646215139442231 | |
| }, | |
| { | |
| "current_steps": 9173, | |
| "loss": 1.3042, | |
| "learning_rate": 0.000634020618556701, | |
| "epoch": 3.6717131474103586 | |
| }, | |
| { | |
| "current_steps": 9237, | |
| "loss": 1.2612, | |
| "learning_rate": 0.0006314432989690721, | |
| "epoch": 3.6972111553784863 | |
| }, | |
| { | |
| "current_steps": 9301, | |
| "loss": 1.2516, | |
| "learning_rate": 0.0006288659793814433, | |
| "epoch": 3.7227091633466136 | |
| }, | |
| { | |
| "current_steps": 9365, | |
| "loss": 1.2572, | |
| "learning_rate": 0.0006262886597938144, | |
| "epoch": 3.748207171314741 | |
| }, | |
| { | |
| "current_steps": 9429, | |
| "loss": 1.2525, | |
| "learning_rate": 0.0006237113402061856, | |
| "epoch": 3.7737051792828686 | |
| }, | |
| { | |
| "current_steps": 9493, | |
| "loss": 1.2509, | |
| "learning_rate": 0.0006211340206185567, | |
| "epoch": 3.799203187250996 | |
| }, | |
| { | |
| "current_steps": 9557, | |
| "loss": 1.2467, | |
| "learning_rate": 0.0006185567010309279, | |
| "epoch": 3.8247011952191237 | |
| }, | |
| { | |
| "current_steps": 9621, | |
| "loss": 1.2375, | |
| "learning_rate": 0.000615979381443299, | |
| "epoch": 3.850199203187251 | |
| }, | |
| { | |
| "current_steps": 9685, | |
| "loss": 1.2406, | |
| "learning_rate": 0.0006134020618556702, | |
| "epoch": 3.8756972111553782 | |
| }, | |
| { | |
| "current_steps": 9749, | |
| "loss": 1.2517, | |
| "learning_rate": 0.0006108247422680413, | |
| "epoch": 3.901195219123506 | |
| }, | |
| { | |
| "current_steps": 9813, | |
| "loss": 1.2533, | |
| "learning_rate": 0.0006082474226804123, | |
| "epoch": 3.9266932270916337 | |
| }, | |
| { | |
| "current_steps": 9877, | |
| "loss": 1.256, | |
| "learning_rate": 0.0006056701030927834, | |
| "epoch": 3.952191235059761 | |
| }, | |
| { | |
| "current_steps": 9941, | |
| "loss": 1.2826, | |
| "learning_rate": 0.0006030927835051546, | |
| "epoch": 3.9776892430278883 | |
| }, | |
| { | |
| "current_steps": 9991, | |
| "loss": 1.226, | |
| "learning_rate": 0.0006005154639175257, | |
| "epoch": 4.003187250996016 | |
| }, | |
| { | |
| "current_steps": 10055, | |
| "loss": 0.9512, | |
| "learning_rate": 0.0005979381443298969, | |
| "epoch": 4.028685258964144 | |
| }, | |
| { | |
| "current_steps": 10055, | |
| "loss": 0.9512, | |
| "learning_rate": 0.0005979381443298969, | |
| "epoch": 4.028685258964144 | |
| } | |
| ] |