| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 99.48249027237354, | |
| "eval_steps": 500, | |
| "global_step": 19200, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.9961089494163424, | |
| "grad_norm": 0.9873842597007751, | |
| "learning_rate": 9.901562500000001e-06, | |
| "loss": 2.5953, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 1.9909208819714657, | |
| "grad_norm": 1.439842700958252, | |
| "learning_rate": 9.8015625e-06, | |
| "loss": 2.462, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 2.985732814526589, | |
| "grad_norm": 2.307729482650757, | |
| "learning_rate": 9.701562500000001e-06, | |
| "loss": 2.3659, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 3.980544747081712, | |
| "grad_norm": 4.006823539733887, | |
| "learning_rate": 9.6015625e-06, | |
| "loss": 2.3475, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 4.975356679636835, | |
| "grad_norm": 2.971815824508667, | |
| "learning_rate": 9.501562500000001e-06, | |
| "loss": 2.3162, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 5.970168612191959, | |
| "grad_norm": 3.6012327671051025, | |
| "learning_rate": 9.4015625e-06, | |
| "loss": 2.3049, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 6.964980544747082, | |
| "grad_norm": 3.803527355194092, | |
| "learning_rate": 9.301562500000002e-06, | |
| "loss": 2.2733, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 7.959792477302205, | |
| "grad_norm": 3.0761234760284424, | |
| "learning_rate": 9.201562500000001e-06, | |
| "loss": 2.2576, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 8.954604409857328, | |
| "grad_norm": 2.6708500385284424, | |
| "learning_rate": 9.1015625e-06, | |
| "loss": 2.2367, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 9.949416342412452, | |
| "grad_norm": 3.482462167739868, | |
| "learning_rate": 9.001562500000001e-06, | |
| "loss": 2.176, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 10.944228274967575, | |
| "grad_norm": 2.826169967651367, | |
| "learning_rate": 8.902083333333334e-06, | |
| "loss": 2.2051, | |
| "step": 2112 | |
| }, | |
| { | |
| "epoch": 11.939040207522698, | |
| "grad_norm": 3.164243698120117, | |
| "learning_rate": 8.802083333333335e-06, | |
| "loss": 2.1874, | |
| "step": 2304 | |
| }, | |
| { | |
| "epoch": 12.933852140077821, | |
| "grad_norm": 4.725254058837891, | |
| "learning_rate": 8.702083333333334e-06, | |
| "loss": 2.1809, | |
| "step": 2496 | |
| }, | |
| { | |
| "epoch": 13.928664072632944, | |
| "grad_norm": 3.7806782722473145, | |
| "learning_rate": 8.602083333333333e-06, | |
| "loss": 2.16, | |
| "step": 2688 | |
| }, | |
| { | |
| "epoch": 14.923476005188068, | |
| "grad_norm": 3.735746145248413, | |
| "learning_rate": 8.502083333333334e-06, | |
| "loss": 2.1663, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 15.918287937743191, | |
| "grad_norm": 3.401533365249634, | |
| "learning_rate": 8.402083333333334e-06, | |
| "loss": 2.1353, | |
| "step": 3072 | |
| }, | |
| { | |
| "epoch": 16.913099870298314, | |
| "grad_norm": 3.1076550483703613, | |
| "learning_rate": 8.302083333333335e-06, | |
| "loss": 2.1277, | |
| "step": 3264 | |
| }, | |
| { | |
| "epoch": 17.907911802853437, | |
| "grad_norm": 3.2938179969787598, | |
| "learning_rate": 8.202083333333334e-06, | |
| "loss": 2.1165, | |
| "step": 3456 | |
| }, | |
| { | |
| "epoch": 18.90272373540856, | |
| "grad_norm": 3.4992620944976807, | |
| "learning_rate": 8.102083333333333e-06, | |
| "loss": 2.1407, | |
| "step": 3648 | |
| }, | |
| { | |
| "epoch": 19.897535667963684, | |
| "grad_norm": 5.144476413726807, | |
| "learning_rate": 8.002083333333334e-06, | |
| "loss": 2.1124, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 20.892347600518807, | |
| "grad_norm": 3.369130849838257, | |
| "learning_rate": 7.902083333333333e-06, | |
| "loss": 2.0994, | |
| "step": 4032 | |
| }, | |
| { | |
| "epoch": 21.88715953307393, | |
| "grad_norm": 2.5219101905822754, | |
| "learning_rate": 7.802604166666668e-06, | |
| "loss": 2.086, | |
| "step": 4224 | |
| }, | |
| { | |
| "epoch": 22.881971465629054, | |
| "grad_norm": 2.416583776473999, | |
| "learning_rate": 7.702604166666667e-06, | |
| "loss": 2.1011, | |
| "step": 4416 | |
| }, | |
| { | |
| "epoch": 23.876783398184177, | |
| "grad_norm": 2.880645513534546, | |
| "learning_rate": 7.602604166666667e-06, | |
| "loss": 2.069, | |
| "step": 4608 | |
| }, | |
| { | |
| "epoch": 24.8715953307393, | |
| "grad_norm": 3.6908700466156006, | |
| "learning_rate": 7.5026041666666675e-06, | |
| "loss": 2.0878, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 25.866407263294423, | |
| "grad_norm": 3.9834940433502197, | |
| "learning_rate": 7.402604166666667e-06, | |
| "loss": 2.0832, | |
| "step": 4992 | |
| }, | |
| { | |
| "epoch": 26.861219195849547, | |
| "grad_norm": 4.434403896331787, | |
| "learning_rate": 7.302604166666667e-06, | |
| "loss": 2.0618, | |
| "step": 5184 | |
| }, | |
| { | |
| "epoch": 27.85603112840467, | |
| "grad_norm": 2.6179683208465576, | |
| "learning_rate": 7.202604166666667e-06, | |
| "loss": 2.0698, | |
| "step": 5376 | |
| }, | |
| { | |
| "epoch": 28.850843060959793, | |
| "grad_norm": 3.7101306915283203, | |
| "learning_rate": 7.102604166666667e-06, | |
| "loss": 2.0561, | |
| "step": 5568 | |
| }, | |
| { | |
| "epoch": 29.845654993514916, | |
| "grad_norm": 2.940573215484619, | |
| "learning_rate": 7.002604166666668e-06, | |
| "loss": 2.0499, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 30.84046692607004, | |
| "grad_norm": 3.1798958778381348, | |
| "learning_rate": 6.902604166666667e-06, | |
| "loss": 2.049, | |
| "step": 5952 | |
| }, | |
| { | |
| "epoch": 31.835278858625163, | |
| "grad_norm": 3.288163423538208, | |
| "learning_rate": 6.802604166666667e-06, | |
| "loss": 2.0326, | |
| "step": 6144 | |
| }, | |
| { | |
| "epoch": 32.83009079118028, | |
| "grad_norm": 2.8764078617095947, | |
| "learning_rate": 6.703125e-06, | |
| "loss": 2.0371, | |
| "step": 6336 | |
| }, | |
| { | |
| "epoch": 33.824902723735406, | |
| "grad_norm": 2.9869441986083984, | |
| "learning_rate": 6.603125e-06, | |
| "loss": 2.0611, | |
| "step": 6528 | |
| }, | |
| { | |
| "epoch": 34.81971465629053, | |
| "grad_norm": 3.1720468997955322, | |
| "learning_rate": 6.5031250000000005e-06, | |
| "loss": 2.0541, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 35.81452658884565, | |
| "grad_norm": 3.7449281215667725, | |
| "learning_rate": 6.403125e-06, | |
| "loss": 2.0177, | |
| "step": 6912 | |
| }, | |
| { | |
| "epoch": 36.809338521400775, | |
| "grad_norm": 2.036007881164551, | |
| "learning_rate": 6.303125000000001e-06, | |
| "loss": 2.006, | |
| "step": 7104 | |
| }, | |
| { | |
| "epoch": 37.8041504539559, | |
| "grad_norm": 2.677340507507324, | |
| "learning_rate": 6.203125000000001e-06, | |
| "loss": 2.0329, | |
| "step": 7296 | |
| }, | |
| { | |
| "epoch": 38.79896238651102, | |
| "grad_norm": 2.8975298404693604, | |
| "learning_rate": 6.103125000000001e-06, | |
| "loss": 2.0347, | |
| "step": 7488 | |
| }, | |
| { | |
| "epoch": 39.793774319066145, | |
| "grad_norm": 1.8516769409179688, | |
| "learning_rate": 6.003125000000001e-06, | |
| "loss": 2.0179, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 40.78858625162127, | |
| "grad_norm": 2.984468698501587, | |
| "learning_rate": 5.903125e-06, | |
| "loss": 2.0002, | |
| "step": 7872 | |
| }, | |
| { | |
| "epoch": 41.78339818417639, | |
| "grad_norm": 4.378921031951904, | |
| "learning_rate": 5.8031250000000004e-06, | |
| "loss": 2.0032, | |
| "step": 8064 | |
| }, | |
| { | |
| "epoch": 42.778210116731515, | |
| "grad_norm": 3.600853204727173, | |
| "learning_rate": 5.7031250000000006e-06, | |
| "loss": 2.0131, | |
| "step": 8256 | |
| }, | |
| { | |
| "epoch": 43.77302204928664, | |
| "grad_norm": 3.475604295730591, | |
| "learning_rate": 5.603125000000001e-06, | |
| "loss": 2.0167, | |
| "step": 8448 | |
| }, | |
| { | |
| "epoch": 44.76783398184176, | |
| "grad_norm": 3.7388381958007812, | |
| "learning_rate": 5.503125e-06, | |
| "loss": 2.0286, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 45.762645914396884, | |
| "grad_norm": 3.0160775184631348, | |
| "learning_rate": 5.403125e-06, | |
| "loss": 2.0327, | |
| "step": 8832 | |
| }, | |
| { | |
| "epoch": 46.75745784695201, | |
| "grad_norm": 2.4288928508758545, | |
| "learning_rate": 5.303645833333334e-06, | |
| "loss": 1.9837, | |
| "step": 9024 | |
| }, | |
| { | |
| "epoch": 47.75226977950713, | |
| "grad_norm": 2.6072142124176025, | |
| "learning_rate": 5.203645833333334e-06, | |
| "loss": 2.022, | |
| "step": 9216 | |
| }, | |
| { | |
| "epoch": 48.747081712062254, | |
| "grad_norm": 3.0784308910369873, | |
| "learning_rate": 5.103645833333334e-06, | |
| "loss": 1.9946, | |
| "step": 9408 | |
| }, | |
| { | |
| "epoch": 49.74189364461738, | |
| "grad_norm": 2.552196979522705, | |
| "learning_rate": 5.003645833333334e-06, | |
| "loss": 1.9857, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 50.7367055771725, | |
| "grad_norm": 5.163851261138916, | |
| "learning_rate": 4.903645833333333e-06, | |
| "loss": 2.0038, | |
| "step": 9792 | |
| }, | |
| { | |
| "epoch": 51.731517509727624, | |
| "grad_norm": 3.9303905963897705, | |
| "learning_rate": 4.8036458333333335e-06, | |
| "loss": 2.0327, | |
| "step": 9984 | |
| }, | |
| { | |
| "epoch": 52.72632944228275, | |
| "grad_norm": 4.269293308258057, | |
| "learning_rate": 4.703645833333334e-06, | |
| "loss": 1.978, | |
| "step": 10176 | |
| }, | |
| { | |
| "epoch": 53.72114137483787, | |
| "grad_norm": 3.4755337238311768, | |
| "learning_rate": 4.603645833333334e-06, | |
| "loss": 2.0154, | |
| "step": 10368 | |
| }, | |
| { | |
| "epoch": 54.715953307392994, | |
| "grad_norm": 2.2201080322265625, | |
| "learning_rate": 4.503645833333334e-06, | |
| "loss": 1.9945, | |
| "step": 10560 | |
| }, | |
| { | |
| "epoch": 55.71076523994812, | |
| "grad_norm": 3.3768227100372314, | |
| "learning_rate": 4.403645833333334e-06, | |
| "loss": 1.9764, | |
| "step": 10752 | |
| }, | |
| { | |
| "epoch": 56.70557717250324, | |
| "grad_norm": 2.9863104820251465, | |
| "learning_rate": 4.303645833333334e-06, | |
| "loss": 2.0031, | |
| "step": 10944 | |
| }, | |
| { | |
| "epoch": 57.70038910505836, | |
| "grad_norm": 3.351330280303955, | |
| "learning_rate": 4.203645833333333e-06, | |
| "loss": 1.9915, | |
| "step": 11136 | |
| }, | |
| { | |
| "epoch": 58.69520103761349, | |
| "grad_norm": 2.745793104171753, | |
| "learning_rate": 4.103645833333333e-06, | |
| "loss": 2.0054, | |
| "step": 11328 | |
| }, | |
| { | |
| "epoch": 59.69001297016861, | |
| "grad_norm": 4.333703517913818, | |
| "learning_rate": 4.0036458333333335e-06, | |
| "loss": 1.9629, | |
| "step": 11520 | |
| }, | |
| { | |
| "epoch": 60.68482490272373, | |
| "grad_norm": 3.1686012744903564, | |
| "learning_rate": 3.903645833333334e-06, | |
| "loss": 2.0011, | |
| "step": 11712 | |
| }, | |
| { | |
| "epoch": 61.679636835278856, | |
| "grad_norm": 2.1179568767547607, | |
| "learning_rate": 3.804166666666667e-06, | |
| "loss": 1.9949, | |
| "step": 11904 | |
| }, | |
| { | |
| "epoch": 62.67444876783398, | |
| "grad_norm": 2.4998276233673096, | |
| "learning_rate": 3.704166666666667e-06, | |
| "loss": 1.9791, | |
| "step": 12096 | |
| }, | |
| { | |
| "epoch": 63.6692607003891, | |
| "grad_norm": 3.710357189178467, | |
| "learning_rate": 3.6041666666666667e-06, | |
| "loss": 1.9757, | |
| "step": 12288 | |
| }, | |
| { | |
| "epoch": 64.66407263294423, | |
| "grad_norm": 3.4601991176605225, | |
| "learning_rate": 3.504166666666667e-06, | |
| "loss": 2.025, | |
| "step": 12480 | |
| }, | |
| { | |
| "epoch": 65.65888456549935, | |
| "grad_norm": 2.626007318496704, | |
| "learning_rate": 3.4041666666666665e-06, | |
| "loss": 1.9755, | |
| "step": 12672 | |
| }, | |
| { | |
| "epoch": 66.65369649805447, | |
| "grad_norm": 3.1454813480377197, | |
| "learning_rate": 3.304166666666667e-06, | |
| "loss": 1.9984, | |
| "step": 12864 | |
| }, | |
| { | |
| "epoch": 67.6485084306096, | |
| "grad_norm": 7.283568859100342, | |
| "learning_rate": 3.204166666666667e-06, | |
| "loss": 1.9776, | |
| "step": 13056 | |
| }, | |
| { | |
| "epoch": 68.64332036316472, | |
| "grad_norm": 2.031538724899292, | |
| "learning_rate": 3.104166666666667e-06, | |
| "loss": 1.9939, | |
| "step": 13248 | |
| }, | |
| { | |
| "epoch": 69.63813229571984, | |
| "grad_norm": 2.370424270629883, | |
| "learning_rate": 3.004166666666667e-06, | |
| "loss": 1.9805, | |
| "step": 13440 | |
| }, | |
| { | |
| "epoch": 70.63294422827497, | |
| "grad_norm": 3.6357874870300293, | |
| "learning_rate": 2.9041666666666667e-06, | |
| "loss": 1.9998, | |
| "step": 13632 | |
| }, | |
| { | |
| "epoch": 71.62775616083009, | |
| "grad_norm": 3.61807918548584, | |
| "learning_rate": 2.8041666666666668e-06, | |
| "loss": 1.9615, | |
| "step": 13824 | |
| }, | |
| { | |
| "epoch": 72.62256809338521, | |
| "grad_norm": 2.440490484237671, | |
| "learning_rate": 2.7041666666666673e-06, | |
| "loss": 1.9715, | |
| "step": 14016 | |
| }, | |
| { | |
| "epoch": 73.61738002594034, | |
| "grad_norm": 2.490007162094116, | |
| "learning_rate": 2.6046875000000006e-06, | |
| "loss": 1.9765, | |
| "step": 14208 | |
| }, | |
| { | |
| "epoch": 74.61219195849546, | |
| "grad_norm": 2.368825674057007, | |
| "learning_rate": 2.5046875000000003e-06, | |
| "loss": 1.9819, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 75.60700389105058, | |
| "grad_norm": 2.9740967750549316, | |
| "learning_rate": 2.4046875000000004e-06, | |
| "loss": 1.9702, | |
| "step": 14592 | |
| }, | |
| { | |
| "epoch": 76.6018158236057, | |
| "grad_norm": 3.215397834777832, | |
| "learning_rate": 2.3046875e-06, | |
| "loss": 1.9913, | |
| "step": 14784 | |
| }, | |
| { | |
| "epoch": 77.59662775616083, | |
| "grad_norm": 2.465824604034424, | |
| "learning_rate": 2.2046875000000002e-06, | |
| "loss": 1.9751, | |
| "step": 14976 | |
| }, | |
| { | |
| "epoch": 78.59143968871595, | |
| "grad_norm": 2.2337405681610107, | |
| "learning_rate": 2.1046875000000003e-06, | |
| "loss": 1.9496, | |
| "step": 15168 | |
| }, | |
| { | |
| "epoch": 79.58625162127107, | |
| "grad_norm": 2.875598907470703, | |
| "learning_rate": 2.0046875e-06, | |
| "loss": 2.0066, | |
| "step": 15360 | |
| }, | |
| { | |
| "epoch": 80.5810635538262, | |
| "grad_norm": 2.8990750312805176, | |
| "learning_rate": 1.9046875000000001e-06, | |
| "loss": 1.9466, | |
| "step": 15552 | |
| }, | |
| { | |
| "epoch": 81.57587548638132, | |
| "grad_norm": 4.734499454498291, | |
| "learning_rate": 1.8046875000000002e-06, | |
| "loss": 1.9636, | |
| "step": 15744 | |
| }, | |
| { | |
| "epoch": 82.57068741893644, | |
| "grad_norm": 2.9932515621185303, | |
| "learning_rate": 1.7046875000000001e-06, | |
| "loss": 1.973, | |
| "step": 15936 | |
| }, | |
| { | |
| "epoch": 83.56549935149157, | |
| "grad_norm": 4.040909767150879, | |
| "learning_rate": 1.6046875e-06, | |
| "loss": 1.9678, | |
| "step": 16128 | |
| }, | |
| { | |
| "epoch": 84.56031128404669, | |
| "grad_norm": 3.0035159587860107, | |
| "learning_rate": 1.5046875000000002e-06, | |
| "loss": 1.9457, | |
| "step": 16320 | |
| }, | |
| { | |
| "epoch": 85.55512321660181, | |
| "grad_norm": 2.5469093322753906, | |
| "learning_rate": 1.4046875e-06, | |
| "loss": 1.9642, | |
| "step": 16512 | |
| }, | |
| { | |
| "epoch": 86.54993514915694, | |
| "grad_norm": 2.850858688354492, | |
| "learning_rate": 1.3046875e-06, | |
| "loss": 1.9638, | |
| "step": 16704 | |
| }, | |
| { | |
| "epoch": 87.54474708171206, | |
| "grad_norm": 2.5728836059570312, | |
| "learning_rate": 1.2046875e-06, | |
| "loss": 1.9681, | |
| "step": 16896 | |
| }, | |
| { | |
| "epoch": 88.53955901426718, | |
| "grad_norm": 1.4205690622329712, | |
| "learning_rate": 1.1046875000000002e-06, | |
| "loss": 1.9585, | |
| "step": 17088 | |
| }, | |
| { | |
| "epoch": 89.5343709468223, | |
| "grad_norm": 1.9591172933578491, | |
| "learning_rate": 1.0046875e-06, | |
| "loss": 1.9894, | |
| "step": 17280 | |
| }, | |
| { | |
| "epoch": 90.52918287937743, | |
| "grad_norm": 3.981717348098755, | |
| "learning_rate": 9.046875000000001e-07, | |
| "loss": 1.9388, | |
| "step": 17472 | |
| }, | |
| { | |
| "epoch": 91.52399481193255, | |
| "grad_norm": 2.348127841949463, | |
| "learning_rate": 8.046875000000001e-07, | |
| "loss": 1.9677, | |
| "step": 17664 | |
| }, | |
| { | |
| "epoch": 92.51880674448768, | |
| "grad_norm": 3.0600669384002686, | |
| "learning_rate": 7.046875e-07, | |
| "loss": 1.9747, | |
| "step": 17856 | |
| }, | |
| { | |
| "epoch": 93.5136186770428, | |
| "grad_norm": 3.865741491317749, | |
| "learning_rate": 6.046875000000001e-07, | |
| "loss": 1.958, | |
| "step": 18048 | |
| }, | |
| { | |
| "epoch": 94.50843060959792, | |
| "grad_norm": 2.755720853805542, | |
| "learning_rate": 5.052083333333334e-07, | |
| "loss": 1.975, | |
| "step": 18240 | |
| }, | |
| { | |
| "epoch": 95.50324254215305, | |
| "grad_norm": 3.4946985244750977, | |
| "learning_rate": 4.0520833333333335e-07, | |
| "loss": 1.9426, | |
| "step": 18432 | |
| }, | |
| { | |
| "epoch": 96.49805447470817, | |
| "grad_norm": 2.993678331375122, | |
| "learning_rate": 3.0520833333333336e-07, | |
| "loss": 1.9497, | |
| "step": 18624 | |
| }, | |
| { | |
| "epoch": 97.49286640726329, | |
| "grad_norm": 3.959327459335327, | |
| "learning_rate": 2.0520833333333334e-07, | |
| "loss": 1.9409, | |
| "step": 18816 | |
| }, | |
| { | |
| "epoch": 98.48767833981842, | |
| "grad_norm": 2.147462844848633, | |
| "learning_rate": 1.0520833333333334e-07, | |
| "loss": 1.9485, | |
| "step": 19008 | |
| }, | |
| { | |
| "epoch": 99.48249027237354, | |
| "grad_norm": 2.728996992111206, | |
| "learning_rate": 5.208333333333334e-09, | |
| "loss": 1.981, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 99.48249027237354, | |
| "step": 19200, | |
| "total_flos": 1.0064407234989773e+18, | |
| "train_loss": 2.0492228651046753, | |
| "train_runtime": 20851.1432, | |
| "train_samples_per_second": 184.834, | |
| "train_steps_per_second": 0.921 | |
| } | |
| ], | |
| "logging_steps": 192, | |
| "max_steps": 19200, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 100, | |
| "save_steps": 6800, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0064407234989773e+18, | |
| "train_batch_size": 50, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |