| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.013256206314774645, | |
| "eval_steps": 500, | |
| "global_step": 110, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00012051096649795132, | |
| "grad_norm": 0.7307866811752319, | |
| "learning_rate": 4e-05, | |
| "loss": 1.2502, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.00024102193299590263, | |
| "grad_norm": 0.7944597601890564, | |
| "learning_rate": 8e-05, | |
| "loss": 1.0923, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.00036153289949385393, | |
| "grad_norm": 0.8116863965988159, | |
| "learning_rate": 0.00012, | |
| "loss": 1.4372, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.00048204386599180526, | |
| "grad_norm": 0.6883746981620789, | |
| "learning_rate": 0.00016, | |
| "loss": 1.2503, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0006025548324897565, | |
| "grad_norm": 0.6956741809844971, | |
| "learning_rate": 0.0002, | |
| "loss": 1.135, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0007230657989877079, | |
| "grad_norm": 0.7852187752723694, | |
| "learning_rate": 0.0001980952380952381, | |
| "loss": 1.0132, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0008435767654856592, | |
| "grad_norm": 0.4692592918872833, | |
| "learning_rate": 0.0001961904761904762, | |
| "loss": 0.7826, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0009640877319836105, | |
| "grad_norm": 0.27623867988586426, | |
| "learning_rate": 0.0001942857142857143, | |
| "loss": 0.664, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0010845986984815619, | |
| "grad_norm": 0.21396474540233612, | |
| "learning_rate": 0.0001923809523809524, | |
| "loss": 0.9179, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.001205109664979513, | |
| "grad_norm": 0.1967506855726242, | |
| "learning_rate": 0.00019047619047619048, | |
| "loss": 0.6711, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0013256206314774645, | |
| "grad_norm": 0.20955657958984375, | |
| "learning_rate": 0.00018857142857142857, | |
| "loss": 0.8331, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.0014461315979754157, | |
| "grad_norm": 0.2680826485157013, | |
| "learning_rate": 0.0001866666666666667, | |
| "loss": 0.8829, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.0015666425644733672, | |
| "grad_norm": 0.25052550435066223, | |
| "learning_rate": 0.00018476190476190478, | |
| "loss": 0.7536, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.0016871535309713184, | |
| "grad_norm": 0.27972114086151123, | |
| "learning_rate": 0.00018285714285714286, | |
| "loss": 0.8129, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.0018076644974692696, | |
| "grad_norm": 0.23484091460704803, | |
| "learning_rate": 0.00018095238095238095, | |
| "loss": 0.8715, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.001928175463967221, | |
| "grad_norm": 0.2122180461883545, | |
| "learning_rate": 0.00017904761904761907, | |
| "loss": 0.9421, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.0020486864304651723, | |
| "grad_norm": 0.19645242393016815, | |
| "learning_rate": 0.00017714285714285713, | |
| "loss": 0.6596, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.0021691973969631237, | |
| "grad_norm": 0.21645572781562805, | |
| "learning_rate": 0.00017523809523809525, | |
| "loss": 0.764, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.002289708363461075, | |
| "grad_norm": 0.15910537540912628, | |
| "learning_rate": 0.00017333333333333334, | |
| "loss": 0.7156, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.002410219329959026, | |
| "grad_norm": 0.1565140336751938, | |
| "learning_rate": 0.00017142857142857143, | |
| "loss": 0.6023, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0025307302964569776, | |
| "grad_norm": 0.17277204990386963, | |
| "learning_rate": 0.00016952380952380954, | |
| "loss": 0.5594, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.002651241262954929, | |
| "grad_norm": 0.17323294281959534, | |
| "learning_rate": 0.00016761904761904763, | |
| "loss": 0.681, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.00277175222945288, | |
| "grad_norm": 0.1539444774389267, | |
| "learning_rate": 0.00016571428571428575, | |
| "loss": 0.7535, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.0028922631959508315, | |
| "grad_norm": 0.16936075687408447, | |
| "learning_rate": 0.0001638095238095238, | |
| "loss": 0.5524, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.003012774162448783, | |
| "grad_norm": 0.1893339455127716, | |
| "learning_rate": 0.00016190476190476192, | |
| "loss": 0.802, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0031332851289467343, | |
| "grad_norm": 0.17078277468681335, | |
| "learning_rate": 0.00016, | |
| "loss": 0.677, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.0032537960954446853, | |
| "grad_norm": 0.1889839768409729, | |
| "learning_rate": 0.0001580952380952381, | |
| "loss": 0.5551, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.003374307061942637, | |
| "grad_norm": 0.2148876190185547, | |
| "learning_rate": 0.0001561904761904762, | |
| "loss": 0.6161, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.0034948180284405882, | |
| "grad_norm": 0.1392691731452942, | |
| "learning_rate": 0.0001542857142857143, | |
| "loss": 0.5348, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.0036153289949385392, | |
| "grad_norm": 0.17458081245422363, | |
| "learning_rate": 0.00015238095238095237, | |
| "loss": 0.7913, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0037358399614364907, | |
| "grad_norm": 0.1562052071094513, | |
| "learning_rate": 0.00015047619047619048, | |
| "loss": 0.8158, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.003856350927934442, | |
| "grad_norm": 0.1435224562883377, | |
| "learning_rate": 0.00014857142857142857, | |
| "loss": 0.7528, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.0039768618944323935, | |
| "grad_norm": 0.14048519730567932, | |
| "learning_rate": 0.00014666666666666666, | |
| "loss": 0.6955, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.0040973728609303445, | |
| "grad_norm": 0.16571789979934692, | |
| "learning_rate": 0.00014476190476190475, | |
| "loss": 0.5537, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.0042178838274282955, | |
| "grad_norm": 0.165692538022995, | |
| "learning_rate": 0.00014285714285714287, | |
| "loss": 0.7134, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.004338394793926247, | |
| "grad_norm": 0.1822883039712906, | |
| "learning_rate": 0.00014095238095238096, | |
| "loss": 0.5432, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.004458905760424198, | |
| "grad_norm": 0.1414850652217865, | |
| "learning_rate": 0.00013904761904761905, | |
| "loss": 0.6703, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.00457941672692215, | |
| "grad_norm": 0.15394528210163116, | |
| "learning_rate": 0.00013714285714285716, | |
| "loss": 0.6561, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.004699927693420101, | |
| "grad_norm": 0.1435491144657135, | |
| "learning_rate": 0.00013523809523809525, | |
| "loss": 0.5644, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.004820438659918052, | |
| "grad_norm": 0.16691423952579498, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 0.7856, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.004940949626416004, | |
| "grad_norm": 0.14211532473564148, | |
| "learning_rate": 0.00013142857142857143, | |
| "loss": 0.6399, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.005061460592913955, | |
| "grad_norm": 0.18083994090557098, | |
| "learning_rate": 0.00012952380952380954, | |
| "loss": 0.715, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.005181971559411906, | |
| "grad_norm": 0.15873770415782928, | |
| "learning_rate": 0.0001276190476190476, | |
| "loss": 0.7614, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.005302482525909858, | |
| "grad_norm": 0.14993314445018768, | |
| "learning_rate": 0.00012571428571428572, | |
| "loss": 0.6105, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.005422993492407809, | |
| "grad_norm": 0.18779931962490082, | |
| "learning_rate": 0.0001238095238095238, | |
| "loss": 1.0716, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.00554350445890576, | |
| "grad_norm": 0.15650784969329834, | |
| "learning_rate": 0.00012190476190476193, | |
| "loss": 0.738, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.005664015425403712, | |
| "grad_norm": 0.1431063711643219, | |
| "learning_rate": 0.00012, | |
| "loss": 0.5219, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.005784526391901663, | |
| "grad_norm": 0.1359708309173584, | |
| "learning_rate": 0.0001180952380952381, | |
| "loss": 0.5886, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.005905037358399614, | |
| "grad_norm": 0.16217978298664093, | |
| "learning_rate": 0.00011619047619047621, | |
| "loss": 0.7634, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.006025548324897566, | |
| "grad_norm": 0.16889767348766327, | |
| "learning_rate": 0.00011428571428571428, | |
| "loss": 0.7717, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.006146059291395517, | |
| "grad_norm": 0.21841812133789062, | |
| "learning_rate": 0.00011238095238095239, | |
| "loss": 0.937, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.006266570257893469, | |
| "grad_norm": 0.17994704842567444, | |
| "learning_rate": 0.00011047619047619049, | |
| "loss": 0.8443, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.00638708122439142, | |
| "grad_norm": 0.15717928111553192, | |
| "learning_rate": 0.00010857142857142856, | |
| "loss": 0.7624, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.006507592190889371, | |
| "grad_norm": 0.16110721230506897, | |
| "learning_rate": 0.00010666666666666667, | |
| "loss": 0.7228, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.0066281031573873226, | |
| "grad_norm": 0.14764989912509918, | |
| "learning_rate": 0.00010476190476190477, | |
| "loss": 0.6782, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.006748614123885274, | |
| "grad_norm": 0.1577727496623993, | |
| "learning_rate": 0.00010285714285714286, | |
| "loss": 0.7367, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.006869125090383225, | |
| "grad_norm": 0.17438825964927673, | |
| "learning_rate": 0.00010095238095238096, | |
| "loss": 0.65, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.0069896360568811764, | |
| "grad_norm": 0.1775740683078766, | |
| "learning_rate": 9.904761904761905e-05, | |
| "loss": 0.7797, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.0071101470233791274, | |
| "grad_norm": 0.18453216552734375, | |
| "learning_rate": 9.714285714285715e-05, | |
| "loss": 0.9153, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.0072306579898770785, | |
| "grad_norm": 0.16022688150405884, | |
| "learning_rate": 9.523809523809524e-05, | |
| "loss": 0.7798, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.00735116895637503, | |
| "grad_norm": 0.16944445669651031, | |
| "learning_rate": 9.333333333333334e-05, | |
| "loss": 0.8193, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.007471679922872981, | |
| "grad_norm": 0.14207735657691956, | |
| "learning_rate": 9.142857142857143e-05, | |
| "loss": 0.5361, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.007592190889370932, | |
| "grad_norm": 0.16854678094387054, | |
| "learning_rate": 8.952380952380953e-05, | |
| "loss": 0.7976, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.007712701855868884, | |
| "grad_norm": 0.17764142155647278, | |
| "learning_rate": 8.761904761904762e-05, | |
| "loss": 0.6938, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.007833212822366835, | |
| "grad_norm": 0.21041354537010193, | |
| "learning_rate": 8.571428571428571e-05, | |
| "loss": 0.8384, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.007953723788864787, | |
| "grad_norm": 0.18576891720294952, | |
| "learning_rate": 8.380952380952382e-05, | |
| "loss": 0.6401, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.008074234755362737, | |
| "grad_norm": 0.20624496042728424, | |
| "learning_rate": 8.19047619047619e-05, | |
| "loss": 0.7563, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.008194745721860689, | |
| "grad_norm": 0.18236589431762695, | |
| "learning_rate": 8e-05, | |
| "loss": 0.748, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.008315256688358641, | |
| "grad_norm": 0.15884153544902802, | |
| "learning_rate": 7.80952380952381e-05, | |
| "loss": 0.649, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.008435767654856591, | |
| "grad_norm": 0.18527762591838837, | |
| "learning_rate": 7.619047619047618e-05, | |
| "loss": 0.5163, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.008556278621354543, | |
| "grad_norm": 0.166184663772583, | |
| "learning_rate": 7.428571428571429e-05, | |
| "loss": 0.7672, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.008676789587852495, | |
| "grad_norm": 0.19784916937351227, | |
| "learning_rate": 7.238095238095238e-05, | |
| "loss": 0.7482, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.008797300554350447, | |
| "grad_norm": 0.16908536851406097, | |
| "learning_rate": 7.047619047619048e-05, | |
| "loss": 0.7461, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.008917811520848397, | |
| "grad_norm": 0.18411517143249512, | |
| "learning_rate": 6.857142857142858e-05, | |
| "loss": 0.5697, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.009038322487346349, | |
| "grad_norm": 0.15351906418800354, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 0.6597, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.0091588334538443, | |
| "grad_norm": 0.17720364034175873, | |
| "learning_rate": 6.476190476190477e-05, | |
| "loss": 0.808, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.00927934442034225, | |
| "grad_norm": 0.18325303494930267, | |
| "learning_rate": 6.285714285714286e-05, | |
| "loss": 0.7917, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.009399855386840203, | |
| "grad_norm": 0.1679506152868271, | |
| "learning_rate": 6.0952380952380964e-05, | |
| "loss": 0.6326, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.009520366353338154, | |
| "grad_norm": 0.19260190427303314, | |
| "learning_rate": 5.904761904761905e-05, | |
| "loss": 0.5601, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.009640877319836105, | |
| "grad_norm": 0.15009605884552002, | |
| "learning_rate": 5.714285714285714e-05, | |
| "loss": 0.6072, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.009761388286334056, | |
| "grad_norm": 0.15776121616363525, | |
| "learning_rate": 5.5238095238095244e-05, | |
| "loss": 0.6753, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.009881899252832008, | |
| "grad_norm": 0.18575388193130493, | |
| "learning_rate": 5.333333333333333e-05, | |
| "loss": 0.6219, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.010002410219329958, | |
| "grad_norm": 0.21978633105754852, | |
| "learning_rate": 5.142857142857143e-05, | |
| "loss": 0.8581, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.01012292118582791, | |
| "grad_norm": 0.1704164743423462, | |
| "learning_rate": 4.9523809523809525e-05, | |
| "loss": 0.6461, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.010243432152325862, | |
| "grad_norm": 0.18057820200920105, | |
| "learning_rate": 4.761904761904762e-05, | |
| "loss": 0.7416, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.010363943118823812, | |
| "grad_norm": 0.15225447714328766, | |
| "learning_rate": 4.5714285714285716e-05, | |
| "loss": 0.4868, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.010484454085321764, | |
| "grad_norm": 0.17193946242332458, | |
| "learning_rate": 4.380952380952381e-05, | |
| "loss": 0.8092, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.010604965051819716, | |
| "grad_norm": 0.194380983710289, | |
| "learning_rate": 4.190476190476191e-05, | |
| "loss": 0.8461, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.010725476018317666, | |
| "grad_norm": 0.2139783650636673, | |
| "learning_rate": 4e-05, | |
| "loss": 0.6548, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.010845986984815618, | |
| "grad_norm": 0.16700893640518188, | |
| "learning_rate": 3.809523809523809e-05, | |
| "loss": 0.5584, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.01096649795131357, | |
| "grad_norm": 0.1971975564956665, | |
| "learning_rate": 3.619047619047619e-05, | |
| "loss": 0.8535, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.01108700891781152, | |
| "grad_norm": 0.19667109847068787, | |
| "learning_rate": 3.428571428571429e-05, | |
| "loss": 0.8635, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.011207519884309472, | |
| "grad_norm": 0.18818983435630798, | |
| "learning_rate": 3.2380952380952386e-05, | |
| "loss": 0.8435, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.011328030850807424, | |
| "grad_norm": 0.16365501284599304, | |
| "learning_rate": 3.0476190476190482e-05, | |
| "loss": 0.6243, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.011448541817305374, | |
| "grad_norm": 0.20358283817768097, | |
| "learning_rate": 2.857142857142857e-05, | |
| "loss": 0.6483, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.011569052783803326, | |
| "grad_norm": 0.17696398496627808, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 0.6057, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.011689563750301278, | |
| "grad_norm": 0.15508583188056946, | |
| "learning_rate": 2.4761904761904762e-05, | |
| "loss": 0.524, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.011810074716799228, | |
| "grad_norm": 0.18458549678325653, | |
| "learning_rate": 2.2857142857142858e-05, | |
| "loss": 0.8364, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.01193058568329718, | |
| "grad_norm": 0.1944003403186798, | |
| "learning_rate": 2.0952380952380954e-05, | |
| "loss": 0.5383, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.012051096649795132, | |
| "grad_norm": 0.4217074513435364, | |
| "learning_rate": 1.9047619047619046e-05, | |
| "loss": 0.6774, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.012171607616293083, | |
| "grad_norm": 0.20350486040115356, | |
| "learning_rate": 1.7142857142857145e-05, | |
| "loss": 0.6871, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.012292118582791034, | |
| "grad_norm": 0.19154471158981323, | |
| "learning_rate": 1.5238095238095241e-05, | |
| "loss": 0.7226, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.012412629549288986, | |
| "grad_norm": 0.17253194749355316, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.7514, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.012533140515786937, | |
| "grad_norm": 0.14699283242225647, | |
| "learning_rate": 1.1428571428571429e-05, | |
| "loss": 0.5358, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.012653651482284888, | |
| "grad_norm": 0.19192050397396088, | |
| "learning_rate": 9.523809523809523e-06, | |
| "loss": 0.9153, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.01277416244878284, | |
| "grad_norm": 0.15646027028560638, | |
| "learning_rate": 7.6190476190476205e-06, | |
| "loss": 0.5182, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.012894673415280791, | |
| "grad_norm": 0.18160918354988098, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 0.5822, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.013015184381778741, | |
| "grad_norm": 0.19203059375286102, | |
| "learning_rate": 3.8095238095238102e-06, | |
| "loss": 0.7678, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.013135695348276693, | |
| "grad_norm": 0.20908264815807343, | |
| "learning_rate": 1.9047619047619051e-06, | |
| "loss": 0.8563, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.013256206314774645, | |
| "grad_norm": 0.16366459429264069, | |
| "learning_rate": 0.0, | |
| "loss": 0.4258, | |
| "step": 110 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 110, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 55, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.103938402981235e+16, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |