| [ | |
| { | |
| "loss": 0.6916, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.02, | |
| "step": 1 | |
| }, | |
| { | |
| "loss": 0.6981, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.04, | |
| "step": 2 | |
| }, | |
| { | |
| "loss": 0.6824, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.05, | |
| "step": 3 | |
| }, | |
| { | |
| "loss": 0.6878, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.07, | |
| "step": 4 | |
| }, | |
| { | |
| "loss": 0.6848, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.09, | |
| "step": 5 | |
| }, | |
| { | |
| "loss": 0.6842, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.11, | |
| "step": 6 | |
| }, | |
| { | |
| "loss": 0.6826, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.12, | |
| "step": 7 | |
| }, | |
| { | |
| "loss": 0.6571, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.14, | |
| "step": 8 | |
| }, | |
| { | |
| "loss": 0.6723, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.16, | |
| "step": 9 | |
| }, | |
| { | |
| "loss": 0.6706, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.18, | |
| "step": 10 | |
| }, | |
| { | |
| "loss": 0.669, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.19, | |
| "step": 11 | |
| }, | |
| { | |
| "loss": 0.6648, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.21, | |
| "step": 12 | |
| }, | |
| { | |
| "loss": 0.6491, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.23, | |
| "step": 13 | |
| }, | |
| { | |
| "loss": 0.6483, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.25, | |
| "step": 14 | |
| }, | |
| { | |
| "loss": 0.6688, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.26, | |
| "step": 15 | |
| }, | |
| { | |
| "loss": 0.6457, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.28, | |
| "step": 16 | |
| }, | |
| { | |
| "loss": 0.6418, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.3, | |
| "step": 17 | |
| }, | |
| { | |
| "loss": 0.642, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.32, | |
| "step": 18 | |
| }, | |
| { | |
| "loss": 0.6052, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.33, | |
| "step": 19 | |
| }, | |
| { | |
| "loss": 0.6353, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.35, | |
| "step": 20 | |
| }, | |
| { | |
| "loss": 0.6342, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.37, | |
| "step": 21 | |
| }, | |
| { | |
| "loss": 0.5882, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.39, | |
| "step": 22 | |
| }, | |
| { | |
| "loss": 0.6016, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.4, | |
| "step": 23 | |
| }, | |
| { | |
| "loss": 0.5912, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.42, | |
| "step": 24 | |
| }, | |
| { | |
| "loss": 0.5357, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.44, | |
| "step": 25 | |
| }, | |
| { | |
| "eval_cooking_loss": 0.5587261915206909, | |
| "eval_cooking_score": -0.18656283617019653, | |
| "eval_cooking_brier_score": 0.18656283617019653, | |
| "eval_cooking_average_probability": 0.5775389671325684, | |
| "eval_cooking_accuracy": 0.86, | |
| "eval_cooking_probabilities": [ | |
| 0.6535859107971191, | |
| 0.7789987921714783, | |
| 0.7278355956077576, | |
| 0.5034284591674805, | |
| 0.5114082098007202, | |
| 0.5213085412979126, | |
| 0.4919251799583435, | |
| 0.5028612613677979, | |
| 0.5094664692878723, | |
| 0.5756015777587891, | |
| 0.649503231048584, | |
| 0.596377432346344, | |
| 0.5078422427177429, | |
| 0.5245868563652039, | |
| 0.49776366353034973, | |
| 0.6764980554580688, | |
| 0.6633638739585876, | |
| 0.7384623885154724, | |
| 0.7947311401367188, | |
| 0.782829761505127, | |
| 0.8216938972473145, | |
| 0.7805530428886414, | |
| 0.763758659362793, | |
| 0.7716487050056458, | |
| 0.5573330521583557, | |
| 0.4893724322319031, | |
| 0.618384838104248, | |
| 0.5045730471611023, | |
| 0.493682861328125, | |
| 0.4950566589832306, | |
| 0.5033178329467773, | |
| 0.5005538463592529, | |
| 0.5081912279129028, | |
| 0.619926929473877, | |
| 0.693365216255188, | |
| 0.5872482061386108, | |
| 0.5683161616325378, | |
| 0.5579834580421448, | |
| 0.5524486303329468, | |
| 0.5865271687507629, | |
| 0.596405029296875, | |
| 0.5350281000137329, | |
| 0.6161436438560486, | |
| 0.6798644661903381, | |
| 0.6344469785690308, | |
| 0.5292232632637024, | |
| 0.5086079239845276, | |
| 0.5822234749794006, | |
| 0.5117786526679993, | |
| 0.519439160823822, | |
| 0.507979154586792, | |
| 0.4805572032928467, | |
| 0.5027401447296143, | |
| 0.496711790561676, | |
| 0.5047993659973145, | |
| 0.50522780418396, | |
| 0.5031599998474121, | |
| 0.5163615345954895, | |
| 0.5120277404785156, | |
| 0.5277899503707886, | |
| 0.7266655564308167, | |
| 0.7472778558731079, | |
| 0.6976069808006287, | |
| 0.5717738270759583, | |
| 0.6468728184700012, | |
| 0.6206830143928528, | |
| 0.49703091382980347, | |
| 0.5049151182174683, | |
| 0.48914843797683716, | |
| 0.5019007325172424, | |
| 0.5251224637031555, | |
| 0.5308592319488525, | |
| 0.5176806449890137, | |
| 0.5524135231971741, | |
| 0.5026034116744995, | |
| 0.5673679113388062, | |
| 0.4874042868614197, | |
| 0.5103644728660583, | |
| 0.49143484234809875, | |
| 0.4760080873966217, | |
| 0.5319715738296509, | |
| 0.6992161273956299, | |
| 0.5555852055549622, | |
| 0.5927191376686096, | |
| 0.4964844882488251, | |
| 0.5331833362579346, | |
| 0.5580875873565674, | |
| 0.5332431197166443, | |
| 0.5607354044914246, | |
| 0.5862356424331665, | |
| 0.5830809473991394, | |
| 0.6106278300285339, | |
| 0.6358562707901001, | |
| 0.5881476402282715, | |
| 0.5585660934448242, | |
| 0.7618038654327393, | |
| 0.5552946329116821, | |
| 0.5067716836929321, | |
| 0.491814523935318, | |
| 0.6965087652206421 | |
| ], | |
| "eval_cooking_runtime": 16.9456, | |
| "eval_cooking_samples_per_second": 5.901, | |
| "eval_cooking_steps_per_second": 0.118, | |
| "epoch": 0.44, | |
| "step": 25 | |
| }, | |
| { | |
| "loss": 0.6283, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.46, | |
| "step": 26 | |
| }, | |
| { | |
| "loss": 0.5924, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.47, | |
| "step": 27 | |
| }, | |
| { | |
| "loss": 0.5645, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.49, | |
| "step": 28 | |
| }, | |
| { | |
| "loss": 0.5242, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.51, | |
| "step": 29 | |
| }, | |
| { | |
| "loss": 0.5376, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.53, | |
| "step": 30 | |
| }, | |
| { | |
| "loss": 0.535, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.54, | |
| "step": 31 | |
| }, | |
| { | |
| "loss": 0.5309, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.56, | |
| "step": 32 | |
| }, | |
| { | |
| "loss": 0.531, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.58, | |
| "step": 33 | |
| }, | |
| { | |
| "loss": 0.4605, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.6, | |
| "step": 34 | |
| }, | |
| { | |
| "loss": 0.4715, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.61, | |
| "step": 35 | |
| }, | |
| { | |
| "loss": 0.4234, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.63, | |
| "step": 36 | |
| }, | |
| { | |
| "loss": 0.4365, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.65, | |
| "step": 37 | |
| }, | |
| { | |
| "loss": 0.4103, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.67, | |
| "step": 38 | |
| }, | |
| { | |
| "loss": 0.3823, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.68, | |
| "step": 39 | |
| }, | |
| { | |
| "loss": 0.4441, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.7, | |
| "step": 40 | |
| }, | |
| { | |
| "loss": 0.3933, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.72, | |
| "step": 41 | |
| }, | |
| { | |
| "loss": 0.2843, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.74, | |
| "step": 42 | |
| }, | |
| { | |
| "loss": 0.369, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.75, | |
| "step": 43 | |
| }, | |
| { | |
| "loss": 0.2645, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.77, | |
| "step": 44 | |
| }, | |
| { | |
| "loss": 0.2849, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.79, | |
| "step": 45 | |
| }, | |
| { | |
| "loss": 0.4064, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.81, | |
| "step": 46 | |
| }, | |
| { | |
| "loss": 0.2997, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.82, | |
| "step": 47 | |
| }, | |
| { | |
| "loss": 0.2671, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.84, | |
| "step": 48 | |
| }, | |
| { | |
| "loss": 0.3314, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.86, | |
| "step": 49 | |
| }, | |
| { | |
| "loss": 0.4259, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.88, | |
| "step": 50 | |
| }, | |
| { | |
| "eval_cooking_loss": 0.24271316826343536, | |
| "eval_cooking_score": -0.07499240338802338, | |
| "eval_cooking_brier_score": 0.07499240338802338, | |
| "eval_cooking_average_probability": 0.8277876973152161, | |
| "eval_cooking_accuracy": 0.9, | |
| "eval_cooking_probabilities": [ | |
| 0.999724805355072, | |
| 0.9999879598617554, | |
| 0.9999728202819824, | |
| 0.7033181190490723, | |
| 0.7430822253227234, | |
| 0.881674587726593, | |
| 0.41176167130470276, | |
| 0.708188533782959, | |
| 0.6328032612800598, | |
| 0.9212954640388489, | |
| 0.9990290403366089, | |
| 0.9299277067184448, | |
| 0.7640402317047119, | |
| 0.3545408844947815, | |
| 0.5027904510498047, | |
| 0.9999113082885742, | |
| 0.9996991157531738, | |
| 0.9999837875366211, | |
| 0.9999978542327881, | |
| 0.9999983310699463, | |
| 0.9999994039535522, | |
| 0.9645881056785583, | |
| 0.9580472707748413, | |
| 0.8383041024208069, | |
| 0.9042266011238098, | |
| 0.34083274006843567, | |
| 0.9612021446228027, | |
| 0.6464914083480835, | |
| 0.5505920648574829, | |
| 0.5640060901641846, | |
| 0.4715419113636017, | |
| 0.5866241455078125, | |
| 0.5923582315444946, | |
| 0.9999804496765137, | |
| 0.9999958276748657, | |
| 0.9999480247497559, | |
| 0.9675708413124084, | |
| 0.9627715349197388, | |
| 0.9368075728416443, | |
| 0.9998013377189636, | |
| 0.9998076558113098, | |
| 0.9965103268623352, | |
| 0.9999927282333374, | |
| 0.9999969005584717, | |
| 0.9999947547912598, | |
| 0.6305108070373535, | |
| 0.9222402572631836, | |
| 0.9166140556335449, | |
| 0.9501265287399292, | |
| 0.8001410961151123, | |
| 0.5643638968467712, | |
| 0.28959962725639343, | |
| 0.46665191650390625, | |
| 0.4726523160934448, | |
| 0.5652157664299011, | |
| 0.6580631136894226, | |
| 0.5187925696372986, | |
| 0.7297425270080566, | |
| 0.6905303001403809, | |
| 0.7946078181266785, | |
| 0.9999961853027344, | |
| 0.9999949932098389, | |
| 0.9999889135360718, | |
| 0.9987756609916687, | |
| 0.9995506405830383, | |
| 0.9991851449012756, | |
| 0.48971259593963623, | |
| 0.5437894463539124, | |
| 0.5442638993263245, | |
| 0.6886961460113525, | |
| 0.8295254707336426, | |
| 0.7418484091758728, | |
| 0.9925927519798279, | |
| 0.9959911704063416, | |
| 0.5952439904212952, | |
| 0.9897077083587646, | |
| 0.621001660823822, | |
| 0.6627683043479919, | |
| 0.42167624831199646, | |
| 0.25209811329841614, | |
| 0.833419680595398, | |
| 0.999998927116394, | |
| 0.9988007545471191, | |
| 0.9999165534973145, | |
| 0.9955934882164001, | |
| 0.9992002844810486, | |
| 0.9901018142700195, | |
| 0.9907991290092468, | |
| 0.9996410608291626, | |
| 0.9999163150787354, | |
| 0.9285746812820435, | |
| 0.9999537467956543, | |
| 0.9999878406524658, | |
| 0.9995831847190857, | |
| 0.999691367149353, | |
| 0.9999963045120239, | |
| 0.9960076808929443, | |
| 0.9942654967308044, | |
| 0.9234359860420227, | |
| 0.9999021291732788 | |
| ], | |
| "eval_cooking_runtime": 16.9232, | |
| "eval_cooking_samples_per_second": 5.909, | |
| "eval_cooking_steps_per_second": 0.118, | |
| "epoch": 0.88, | |
| "step": 50 | |
| }, | |
| { | |
| "loss": 0.179, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.89, | |
| "step": 51 | |
| }, | |
| { | |
| "loss": 0.2263, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.91, | |
| "step": 52 | |
| }, | |
| { | |
| "loss": 0.254, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.93, | |
| "step": 53 | |
| }, | |
| { | |
| "loss": 0.2283, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.95, | |
| "step": 54 | |
| }, | |
| { | |
| "loss": 0.1749, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.96, | |
| "step": 55 | |
| }, | |
| { | |
| "loss": 0.2974, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 0.98, | |
| "step": 56 | |
| }, | |
| { | |
| "loss": 0.2037, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.0, | |
| "step": 57 | |
| }, | |
| { | |
| "loss": 0.2979, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.02, | |
| "step": 58 | |
| }, | |
| { | |
| "loss": 0.1342, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.04, | |
| "step": 59 | |
| }, | |
| { | |
| "loss": 0.1416, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.05, | |
| "step": 60 | |
| }, | |
| { | |
| "loss": 0.1714, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.07, | |
| "step": 61 | |
| }, | |
| { | |
| "loss": 0.1309, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.09, | |
| "step": 62 | |
| }, | |
| { | |
| "loss": 0.21, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.11, | |
| "step": 63 | |
| }, | |
| { | |
| "loss": 0.1542, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.12, | |
| "step": 64 | |
| }, | |
| { | |
| "loss": 0.2992, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.14, | |
| "step": 65 | |
| }, | |
| { | |
| "loss": 0.1631, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.16, | |
| "step": 66 | |
| }, | |
| { | |
| "loss": 0.5016, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.18, | |
| "step": 67 | |
| }, | |
| { | |
| "loss": 0.1565, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.19, | |
| "step": 68 | |
| }, | |
| { | |
| "loss": 0.1252, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.21, | |
| "step": 69 | |
| }, | |
| { | |
| "loss": 0.1232, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.23, | |
| "step": 70 | |
| }, | |
| { | |
| "loss": 0.1189, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.25, | |
| "step": 71 | |
| }, | |
| { | |
| "loss": 0.1427, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.26, | |
| "step": 72 | |
| }, | |
| { | |
| "loss": 0.1295, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.28, | |
| "step": 73 | |
| }, | |
| { | |
| "loss": 0.0967, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.3, | |
| "step": 74 | |
| }, | |
| { | |
| "loss": 0.13, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.32, | |
| "step": 75 | |
| }, | |
| { | |
| "eval_cooking_loss": 0.1628786027431488, | |
| "eval_cooking_score": -0.05112582817673683, | |
| "eval_cooking_brier_score": 0.05112582817673683, | |
| "eval_cooking_average_probability": 0.8956797122955322, | |
| "eval_cooking_accuracy": 0.93, | |
| "eval_cooking_probabilities": [ | |
| 0.9999997615814209, | |
| 1.0, | |
| 1.0, | |
| 0.9542316198348999, | |
| 0.9880340695381165, | |
| 0.9988256096839905, | |
| 0.5675406455993652, | |
| 0.9845830202102661, | |
| 0.9547784328460693, | |
| 0.9989789724349976, | |
| 0.9999997615814209, | |
| 0.9986110925674438, | |
| 0.9338040351867676, | |
| 0.17299044132232666, | |
| 0.45846590399742126, | |
| 0.999994158744812, | |
| 0.9999963045120239, | |
| 0.9999996423721313, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9999418258666992, | |
| 0.99986732006073, | |
| 0.8973928093910217, | |
| 0.9987490177154541, | |
| 0.9780608415603638, | |
| 0.999396562576294, | |
| 0.9193954467773438, | |
| 0.7913320660591125, | |
| 0.723684549331665, | |
| 0.4277496635913849, | |
| 0.7048351168632507, | |
| 0.7073538303375244, | |
| 1.0, | |
| 1.0, | |
| 0.9999998807907104, | |
| 0.9997616410255432, | |
| 0.9998383522033691, | |
| 0.9980823993682861, | |
| 0.9999997615814209, | |
| 0.9999997615814209, | |
| 0.9999886751174927, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.7106478214263916, | |
| 0.9978567957878113, | |
| 0.9251576662063599, | |
| 0.9943885207176208, | |
| 0.9633152484893799, | |
| 0.7126079201698303, | |
| 0.15048664808273315, | |
| 0.8353597521781921, | |
| 0.5454015731811523, | |
| 0.8950506448745728, | |
| 0.7718052268028259, | |
| 0.5149663090705872, | |
| 0.9856672883033752, | |
| 0.998984158039093, | |
| 0.9991474151611328, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9999971389770508, | |
| 0.9999994039535522, | |
| 0.9999972581863403, | |
| 0.7013386487960815, | |
| 0.2816448211669922, | |
| 0.728237509727478, | |
| 0.7397965788841248, | |
| 0.9191795587539673, | |
| 0.886256754398346, | |
| 0.9999814033508301, | |
| 0.9999879598617554, | |
| 0.8164882659912109, | |
| 0.999995231628418, | |
| 0.962520182132721, | |
| 0.9958972930908203, | |
| 0.4359956681728363, | |
| 0.15542198717594147, | |
| 0.7957375049591064, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9998841285705566, | |
| 0.9999792575836182, | |
| 0.9998303651809692, | |
| 0.9999983310699463, | |
| 0.9999995231628418, | |
| 0.9999998807907104, | |
| 0.9967737793922424, | |
| 0.9999997615814209, | |
| 0.9999998807907104, | |
| 0.9999998807907104, | |
| 1.0, | |
| 1.0, | |
| 0.999854564666748, | |
| 0.9992210865020752, | |
| 0.9968494772911072, | |
| 1.0 | |
| ], | |
| "eval_cooking_runtime": 16.9537, | |
| "eval_cooking_samples_per_second": 5.898, | |
| "eval_cooking_steps_per_second": 0.118, | |
| "epoch": 1.32, | |
| "step": 75 | |
| }, | |
| { | |
| "loss": 0.1269, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.33, | |
| "step": 76 | |
| }, | |
| { | |
| "loss": 0.1392, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.35, | |
| "step": 77 | |
| }, | |
| { | |
| "loss": 0.1445, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.37, | |
| "step": 78 | |
| }, | |
| { | |
| "loss": 0.1587, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.39, | |
| "step": 79 | |
| }, | |
| { | |
| "loss": 0.1955, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.4, | |
| "step": 80 | |
| }, | |
| { | |
| "loss": 0.1407, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.42, | |
| "step": 81 | |
| }, | |
| { | |
| "loss": 0.1618, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.44, | |
| "step": 82 | |
| }, | |
| { | |
| "loss": 0.1103, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.46, | |
| "step": 83 | |
| }, | |
| { | |
| "loss": 0.1269, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.47, | |
| "step": 84 | |
| }, | |
| { | |
| "loss": 0.1417, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.49, | |
| "step": 85 | |
| }, | |
| { | |
| "loss": 0.1295, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.51, | |
| "step": 86 | |
| }, | |
| { | |
| "loss": 0.1235, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.53, | |
| "step": 87 | |
| }, | |
| { | |
| "loss": 0.128, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.54, | |
| "step": 88 | |
| }, | |
| { | |
| "loss": 0.3058, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.56, | |
| "step": 89 | |
| }, | |
| { | |
| "loss": 0.1572, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.58, | |
| "step": 90 | |
| }, | |
| { | |
| "loss": 0.1179, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.6, | |
| "step": 91 | |
| }, | |
| { | |
| "loss": 0.1117, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.61, | |
| "step": 92 | |
| }, | |
| { | |
| "loss": 0.1062, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.63, | |
| "step": 93 | |
| }, | |
| { | |
| "loss": 0.0737, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.65, | |
| "step": 94 | |
| }, | |
| { | |
| "loss": 0.1918, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.67, | |
| "step": 95 | |
| }, | |
| { | |
| "loss": 0.1854, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.68, | |
| "step": 96 | |
| }, | |
| { | |
| "loss": 0.0775, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.7, | |
| "step": 97 | |
| }, | |
| { | |
| "loss": 0.0842, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.72, | |
| "step": 98 | |
| }, | |
| { | |
| "loss": 0.0614, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.74, | |
| "step": 99 | |
| }, | |
| { | |
| "loss": 0.0925, | |
| "learning_rate": 7.2e-05, | |
| "epoch": 1.75, | |
| "step": 100 | |
| }, | |
| { | |
| "eval_cooking_loss": 0.16701044142246246, | |
| "eval_cooking_score": -0.049089040607213974, | |
| "eval_cooking_brier_score": 0.049089040607213974, | |
| "eval_cooking_average_probability": 0.9079131484031677, | |
| "eval_cooking_accuracy": 0.93, | |
| "eval_cooking_probabilities": [ | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9960379600524902, | |
| 0.9952751398086548, | |
| 0.9999885559082031, | |
| 0.8473833799362183, | |
| 0.9993119239807129, | |
| 0.9969514608383179, | |
| 0.999823272228241, | |
| 1.0, | |
| 0.9998192191123962, | |
| 0.9292259812355042, | |
| 0.37933260202407837, | |
| 0.36354097723960876, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9999985694885254, | |
| 0.9999842643737793, | |
| 0.8276720643043518, | |
| 0.999998927116394, | |
| 0.9999539852142334, | |
| 0.9999990463256836, | |
| 0.9115281105041504, | |
| 0.7917703986167908, | |
| 0.7387007474899292, | |
| 0.37488386034965515, | |
| 0.8065743446350098, | |
| 0.6968543529510498, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9999852180480957, | |
| 0.9999778270721436, | |
| 0.9997331500053406, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.631415605545044, | |
| 0.9998106360435486, | |
| 0.9700195789337158, | |
| 0.9995588660240173, | |
| 0.9897929430007935, | |
| 0.7991155385971069, | |
| 0.016083385795354843, | |
| 0.7816288471221924, | |
| 0.505143940448761, | |
| 0.9806777834892273, | |
| 0.926846444606781, | |
| 0.6193849444389343, | |
| 0.9996283054351807, | |
| 0.999963641166687, | |
| 0.9999926090240479, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.8756850957870483, | |
| 0.20660391449928284, | |
| 0.8607563972473145, | |
| 0.7930099368095398, | |
| 0.9620106816291809, | |
| 0.9651699066162109, | |
| 1.0, | |
| 1.0, | |
| 0.7446385622024536, | |
| 1.0, | |
| 0.9957782030105591, | |
| 0.9999175071716309, | |
| 0.42324966192245483, | |
| 0.15068158507347107, | |
| 0.9407676458358765, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9999991655349731, | |
| 0.9999998807907104, | |
| 0.9999990463256836, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9996916055679321, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 1.0, | |
| 0.9999998807907104, | |
| 0.9999940395355225, | |
| 0.9999938011169434, | |
| 1.0 | |
| ], | |
| "eval_cooking_runtime": 16.9203, | |
| "eval_cooking_samples_per_second": 5.91, | |
| "eval_cooking_steps_per_second": 0.118, | |
| "epoch": 1.75, | |
| "step": 100 | |
| }, | |
| { | |
| "train_runtime": 1392.5305, | |
| "train_samples_per_second": 2.298, | |
| "train_steps_per_second": 0.072, | |
| "total_flos": 0.0, | |
| "train_loss": 0.3521362992748618, | |
| "epoch": 1.75, | |
| "step": 100 | |
| } | |
| ] |