File size: 16,600 Bytes
049bce0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
{
  "best_metric": 0.57263505,
  "best_model_checkpoint": "/export/home2/zli/kc/mm_rag/Qwen2.5-14B-Instruct_lora/checkpoint-256",
  "epoch": 0.9978075517661389,
  "eval_steps": 100,
  "global_step": 256,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.00389768574908648,
      "grad_norm": 0.3756321966648102,
      "learning_rate": 7.692307692307694e-06,
      "loss": 0.760678231716156,
      "memory(GiB)": 64.79,
      "step": 1,
      "token_acc": 0.794377180381695,
      "train_speed(iter/s)": 0.008942
    },
    {
      "epoch": 0.0194884287454324,
      "grad_norm": 0.4006075859069824,
      "learning_rate": 3.846153846153846e-05,
      "loss": 0.8511521816253662,
      "memory(GiB)": 69.21,
      "step": 5,
      "token_acc": 0.7628462851792697,
      "train_speed(iter/s)": 0.009262
    },
    {
      "epoch": 0.0389768574908648,
      "grad_norm": 0.40541699528694153,
      "learning_rate": 7.692307692307693e-05,
      "loss": 0.8799286842346191,
      "memory(GiB)": 69.21,
      "step": 10,
      "token_acc": 0.755640420079393,
      "train_speed(iter/s)": 0.009459
    },
    {
      "epoch": 0.058465286236297195,
      "grad_norm": 0.28694042563438416,
      "learning_rate": 9.998328666948438e-05,
      "loss": 0.7268682479858398,
      "memory(GiB)": 69.21,
      "step": 15,
      "token_acc": 0.7857048253940223,
      "train_speed(iter/s)": 0.009584
    },
    {
      "epoch": 0.0779537149817296,
      "grad_norm": 0.29239076375961304,
      "learning_rate": 9.979538999730047e-05,
      "loss": 0.6990458011627197,
      "memory(GiB)": 69.21,
      "step": 20,
      "token_acc": 0.7923689929968607,
      "train_speed(iter/s)": 0.009633
    },
    {
      "epoch": 0.09744214372716199,
      "grad_norm": 0.25992509722709656,
      "learning_rate": 9.939949247384046e-05,
      "loss": 0.7058870792388916,
      "memory(GiB)": 71.47,
      "step": 25,
      "token_acc": 0.7879081525173794,
      "train_speed(iter/s)": 0.009549
    },
    {
      "epoch": 0.11693057247259439,
      "grad_norm": 0.1924816370010376,
      "learning_rate": 9.879724780684519e-05,
      "loss": 0.7370085716247559,
      "memory(GiB)": 71.47,
      "step": 30,
      "token_acc": 0.7818104986383698,
      "train_speed(iter/s)": 0.009632
    },
    {
      "epoch": 0.1364190012180268,
      "grad_norm": 0.2375946342945099,
      "learning_rate": 9.799117163889559e-05,
      "loss": 0.6951110363006592,
      "memory(GiB)": 71.47,
      "step": 35,
      "token_acc": 0.7915360501567398,
      "train_speed(iter/s)": 0.009685
    },
    {
      "epoch": 0.1559074299634592,
      "grad_norm": 0.19938509166240692,
      "learning_rate": 9.698463103929542e-05,
      "loss": 0.634189510345459,
      "memory(GiB)": 71.47,
      "step": 40,
      "token_acc": 0.8057028291378926,
      "train_speed(iter/s)": 0.009727
    },
    {
      "epoch": 0.1753958587088916,
      "grad_norm": 0.2671877145767212,
      "learning_rate": 9.57818304394503e-05,
      "loss": 0.6699491500854492,
      "memory(GiB)": 71.47,
      "step": 45,
      "token_acc": 0.8010264675941635,
      "train_speed(iter/s)": 0.009776
    },
    {
      "epoch": 0.19488428745432398,
      "grad_norm": 0.3463566303253174,
      "learning_rate": 9.438779407049281e-05,
      "loss": 0.6760294437408447,
      "memory(GiB)": 71.47,
      "step": 50,
      "token_acc": 0.793778801843318,
      "train_speed(iter/s)": 0.009808
    },
    {
      "epoch": 0.2143727161997564,
      "grad_norm": 0.22926518321037292,
      "learning_rate": 9.280834497651334e-05,
      "loss": 0.6952155113220215,
      "memory(GiB)": 71.47,
      "step": 55,
      "token_acc": 0.7856186209763081,
      "train_speed(iter/s)": 0.009819
    },
    {
      "epoch": 0.23386114494518878,
      "grad_norm": 0.26200923323631287,
      "learning_rate": 9.105008069106093e-05,
      "loss": 0.6922572135925293,
      "memory(GiB)": 71.47,
      "step": 60,
      "token_acc": 0.784200469035004,
      "train_speed(iter/s)": 0.009839
    },
    {
      "epoch": 0.25334957369062117,
      "grad_norm": 0.2608520984649658,
      "learning_rate": 8.912034567851599e-05,
      "loss": 0.6162422180175782,
      "memory(GiB)": 71.47,
      "step": 65,
      "token_acc": 0.8122029543994862,
      "train_speed(iter/s)": 0.009864
    },
    {
      "epoch": 0.2728380024360536,
      "grad_norm": 0.23958414793014526,
      "learning_rate": 8.702720065545024e-05,
      "loss": 0.6201987266540527,
      "memory(GiB)": 71.47,
      "step": 70,
      "token_acc": 0.807689317368994,
      "train_speed(iter/s)": 0.00982
    },
    {
      "epoch": 0.292326431181486,
      "grad_norm": 0.24707700312137604,
      "learning_rate": 8.47793889201221e-05,
      "loss": 0.6751769542694092,
      "memory(GiB)": 71.47,
      "step": 75,
      "token_acc": 0.7933965994464215,
      "train_speed(iter/s)": 0.009843
    },
    {
      "epoch": 0.3118148599269184,
      "grad_norm": 0.280536413192749,
      "learning_rate": 8.238629983075294e-05,
      "loss": 0.6537846088409424,
      "memory(GiB)": 71.47,
      "step": 80,
      "token_acc": 0.800564772326156,
      "train_speed(iter/s)": 0.00983
    },
    {
      "epoch": 0.3313032886723508,
      "grad_norm": 0.29046493768692017,
      "learning_rate": 7.985792958513931e-05,
      "loss": 0.6765205383300781,
      "memory(GiB)": 71.47,
      "step": 85,
      "token_acc": 0.7913814331445457,
      "train_speed(iter/s)": 0.009828
    },
    {
      "epoch": 0.3507917174177832,
      "grad_norm": 0.2879982888698578,
      "learning_rate": 7.720483946542914e-05,
      "loss": 0.6516193866729736,
      "memory(GiB)": 71.47,
      "step": 90,
      "token_acc": 0.7983723394009439,
      "train_speed(iter/s)": 0.009817
    },
    {
      "epoch": 0.3702801461632156,
      "grad_norm": 0.28098541498184204,
      "learning_rate": 7.443811172247821e-05,
      "loss": 0.6455563545227051,
      "memory(GiB)": 71.47,
      "step": 95,
      "token_acc": 0.8004983012457532,
      "train_speed(iter/s)": 0.009833
    },
    {
      "epoch": 0.38976857490864797,
      "grad_norm": 0.2407151460647583,
      "learning_rate": 7.156930328406268e-05,
      "loss": 0.6356846332550049,
      "memory(GiB)": 71.47,
      "step": 100,
      "token_acc": 0.8042028604018602,
      "train_speed(iter/s)": 0.009804
    },
    {
      "epoch": 0.38976857490864797,
      "eval_loss": 0.598721981048584,
      "eval_runtime": 15.1098,
      "eval_samples_per_second": 0.265,
      "eval_steps_per_second": 0.132,
      "step": 100
    },
    {
      "epoch": 0.4092570036540804,
      "grad_norm": 0.2836802899837494,
      "learning_rate": 6.861039748031351e-05,
      "loss": 0.6598537921905517,
      "memory(GiB)": 77.63,
      "step": 105,
      "token_acc": 0.7957189390414147,
      "train_speed(iter/s)": 0.009789
    },
    {
      "epoch": 0.4287454323995128,
      "grad_norm": 0.26613113284111023,
      "learning_rate": 6.557375398802123e-05,
      "loss": 0.6290350437164307,
      "memory(GiB)": 77.63,
      "step": 110,
      "token_acc": 0.8080845540515483,
      "train_speed(iter/s)": 0.009743
    },
    {
      "epoch": 0.4482338611449452,
      "grad_norm": 0.3952692747116089,
      "learning_rate": 6.247205720289907e-05,
      "loss": 0.6128222942352295,
      "memory(GiB)": 77.63,
      "step": 115,
      "token_acc": 0.8121368342872122,
      "train_speed(iter/s)": 0.009719
    },
    {
      "epoch": 0.46772228989037756,
      "grad_norm": 0.29315847158432007,
      "learning_rate": 5.9318263255459116e-05,
      "loss": 0.6240291595458984,
      "memory(GiB)": 77.63,
      "step": 120,
      "token_acc": 0.8027493895269965,
      "train_speed(iter/s)": 0.009694
    },
    {
      "epoch": 0.48721071863581,
      "grad_norm": 0.2671997547149658,
      "learning_rate": 5.6125545891822274e-05,
      "loss": 0.6649426460266114,
      "memory(GiB)": 77.63,
      "step": 125,
      "token_acc": 0.7922591927838373,
      "train_speed(iter/s)": 0.009676
    },
    {
      "epoch": 0.5066991473812423,
      "grad_norm": 0.26974523067474365,
      "learning_rate": 5.290724144552379e-05,
      "loss": 0.6172929763793945,
      "memory(GiB)": 77.63,
      "step": 130,
      "token_acc": 0.8085726351351351,
      "train_speed(iter/s)": 0.009667
    },
    {
      "epoch": 0.5261875761266748,
      "grad_norm": 0.2798590660095215,
      "learning_rate": 4.967679313017303e-05,
      "loss": 0.6145929336547852,
      "memory(GiB)": 77.63,
      "step": 135,
      "token_acc": 0.8118081180811808,
      "train_speed(iter/s)": 0.009683
    },
    {
      "epoch": 0.5456760048721072,
      "grad_norm": 0.27218812704086304,
      "learning_rate": 4.6447694885663514e-05,
      "loss": 0.618541955947876,
      "memory(GiB)": 77.63,
      "step": 140,
      "token_acc": 0.8050674649659422,
      "train_speed(iter/s)": 0.009706
    },
    {
      "epoch": 0.5651644336175395,
      "grad_norm": 0.29498329758644104,
      "learning_rate": 4.323343501249346e-05,
      "loss": 0.6341890335083008,
      "memory(GiB)": 77.63,
      "step": 145,
      "token_acc": 0.801550316061459,
      "train_speed(iter/s)": 0.009735
    },
    {
      "epoch": 0.584652862362972,
      "grad_norm": 0.2799496352672577,
      "learning_rate": 4.004743982964298e-05,
      "loss": 0.6000133037567139,
      "memory(GiB)": 77.63,
      "step": 150,
      "token_acc": 0.8125188083057479,
      "train_speed(iter/s)": 0.009742
    },
    {
      "epoch": 0.6041412911084044,
      "grad_norm": 0.28971585631370544,
      "learning_rate": 3.6903017591354706e-05,
      "loss": 0.6476753711700439,
      "memory(GiB)": 77.63,
      "step": 155,
      "token_acc": 0.8055159053833605,
      "train_speed(iter/s)": 0.009734
    },
    {
      "epoch": 0.6236297198538368,
      "grad_norm": 0.2888241708278656,
      "learning_rate": 3.381330289708396e-05,
      "loss": 0.6604040622711181,
      "memory(GiB)": 77.63,
      "step": 160,
      "token_acc": 0.7972270363951474,
      "train_speed(iter/s)": 0.009743
    },
    {
      "epoch": 0.6431181485992692,
      "grad_norm": 0.31141144037246704,
      "learning_rate": 3.079120182682412e-05,
      "loss": 0.5943418025970459,
      "memory(GiB)": 77.63,
      "step": 165,
      "token_acc": 0.8125655328607977,
      "train_speed(iter/s)": 0.009733
    },
    {
      "epoch": 0.6626065773447016,
      "grad_norm": 0.28593260049819946,
      "learning_rate": 2.7849338030983257e-05,
      "loss": 0.6077968120574951,
      "memory(GiB)": 77.63,
      "step": 170,
      "token_acc": 0.811592125382263,
      "train_speed(iter/s)": 0.009731
    },
    {
      "epoch": 0.682095006090134,
      "grad_norm": 0.3223704993724823,
      "learning_rate": 2.500000000000001e-05,
      "loss": 0.6201191425323487,
      "memory(GiB)": 77.63,
      "step": 175,
      "token_acc": 0.806087783666259,
      "train_speed(iter/s)": 0.009728
    },
    {
      "epoch": 0.7015834348355664,
      "grad_norm": 0.3051086962223053,
      "learning_rate": 2.225508973396016e-05,
      "loss": 0.6003672122955322,
      "memory(GiB)": 77.63,
      "step": 180,
      "token_acc": 0.8096822043421585,
      "train_speed(iter/s)": 0.009738
    },
    {
      "epoch": 0.7210718635809987,
      "grad_norm": 0.33248814940452576,
      "learning_rate": 1.9626073026625818e-05,
      "loss": 0.6649184226989746,
      "memory(GiB)": 77.63,
      "step": 185,
      "token_acc": 0.7927942418271744,
      "train_speed(iter/s)": 0.009754
    },
    {
      "epoch": 0.7405602923264312,
      "grad_norm": 0.32766395807266235,
      "learning_rate": 1.7123931571546827e-05,
      "loss": 0.620220422744751,
      "memory(GiB)": 77.63,
      "step": 190,
      "token_acc": 0.8047315233083407,
      "train_speed(iter/s)": 0.009761
    },
    {
      "epoch": 0.7600487210718636,
      "grad_norm": 0.3414992392063141,
      "learning_rate": 1.4759117090312197e-05,
      "loss": 0.622031831741333,
      "memory(GiB)": 77.63,
      "step": 195,
      "token_acc": 0.807596541264883,
      "train_speed(iter/s)": 0.009765
    },
    {
      "epoch": 0.7795371498172959,
      "grad_norm": 0.32977092266082764,
      "learning_rate": 1.25415076745532e-05,
      "loss": 0.6251283168792725,
      "memory(GiB)": 77.63,
      "step": 200,
      "token_acc": 0.8072234762979684,
      "train_speed(iter/s)": 0.009771
    },
    {
      "epoch": 0.7795371498172959,
      "eval_loss": 0.5760576725006104,
      "eval_runtime": 14.2053,
      "eval_samples_per_second": 0.282,
      "eval_steps_per_second": 0.141,
      "step": 200
    },
    {
      "epoch": 0.7990255785627284,
      "grad_norm": 0.2861402630805969,
      "learning_rate": 1.0480366524062042e-05,
      "loss": 0.5975393772125244,
      "memory(GiB)": 77.63,
      "step": 205,
      "token_acc": 0.8152735624641129,
      "train_speed(iter/s)": 0.009771
    },
    {
      "epoch": 0.8185140073081608,
      "grad_norm": 0.30515122413635254,
      "learning_rate": 8.584303253381847e-06,
      "loss": 0.6246855735778809,
      "memory(GiB)": 77.63,
      "step": 210,
      "token_acc": 0.8046837508398119,
      "train_speed(iter/s)": 0.009789
    },
    {
      "epoch": 0.8380024360535931,
      "grad_norm": 0.3188154101371765,
      "learning_rate": 6.861237928494579e-06,
      "loss": 0.5521413803100585,
      "memory(GiB)": 77.63,
      "step": 215,
      "token_acc": 0.8276072283910298,
      "train_speed(iter/s)": 0.009786
    },
    {
      "epoch": 0.8574908647990256,
      "grad_norm": 0.3368200361728668,
      "learning_rate": 5.318367983829392e-06,
      "loss": 0.5954047679901123,
      "memory(GiB)": 77.63,
      "step": 220,
      "token_acc": 0.8130686517783292,
      "train_speed(iter/s)": 0.009788
    },
    {
      "epoch": 0.876979293544458,
      "grad_norm": 0.3377918601036072,
      "learning_rate": 3.962138157783085e-06,
      "loss": 0.6312428951263428,
      "memory(GiB)": 77.63,
      "step": 225,
      "token_acc": 0.8009544725061238,
      "train_speed(iter/s)": 0.009783
    },
    {
      "epoch": 0.8964677222898904,
      "grad_norm": 0.34851112961769104,
      "learning_rate": 2.798213572335001e-06,
      "loss": 0.5601920127868653,
      "memory(GiB)": 77.63,
      "step": 230,
      "token_acc": 0.8258236434108527,
      "train_speed(iter/s)": 0.009778
    },
    {
      "epoch": 0.9159561510353228,
      "grad_norm": 0.3354482650756836,
      "learning_rate": 1.8314560692059835e-06,
      "loss": 0.5771200656890869,
      "memory(GiB)": 77.63,
      "step": 235,
      "token_acc": 0.8201089968594125,
      "train_speed(iter/s)": 0.009786
    },
    {
      "epoch": 0.9354445797807551,
      "grad_norm": 0.29914259910583496,
      "learning_rate": 1.0659039014077944e-06,
      "loss": 0.6100361347198486,
      "memory(GiB)": 77.63,
      "step": 240,
      "token_acc": 0.8085165989634403,
      "train_speed(iter/s)": 0.009798
    },
    {
      "epoch": 0.9549330085261876,
      "grad_norm": 0.36409541964530945,
      "learning_rate": 5.047548650136513e-07,
      "loss": 0.6194799423217774,
      "memory(GiB)": 77.63,
      "step": 245,
      "token_acc": 0.807660813392776,
      "train_speed(iter/s)": 0.009801
    },
    {
      "epoch": 0.97442143727162,
      "grad_norm": 0.40651243925094604,
      "learning_rate": 1.503529416103988e-07,
      "loss": 0.5671232223510743,
      "memory(GiB)": 77.63,
      "step": 250,
      "token_acc": 0.8216745683871545,
      "train_speed(iter/s)": 0.009809
    },
    {
      "epoch": 0.9939098660170523,
      "grad_norm": 0.3596270978450775,
      "learning_rate": 4.178507228136397e-09,
      "loss": 0.6255767822265625,
      "memory(GiB)": 77.63,
      "step": 255,
      "token_acc": 0.8013068013068013,
      "train_speed(iter/s)": 0.009788
    },
    {
      "epoch": 0.9978075517661389,
      "eval_loss": 0.5726350545883179,
      "eval_runtime": 14.0103,
      "eval_samples_per_second": 0.286,
      "eval_steps_per_second": 0.143,
      "step": 256
    }
  ],
  "logging_steps": 5,
  "max_steps": 256,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 50,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 1.3517329026916086e+18,
  "train_batch_size": 1,
  "trial_name": null,
  "trial_params": null
}