diff --git "a/last-checkpoint/trainer_state.json" "b/last-checkpoint/trainer_state.json" --- "a/last-checkpoint/trainer_state.json" +++ "b/last-checkpoint/trainer_state.json" @@ -1,6527 +1,5360 @@ { "best_metric": null, "best_model_checkpoint": null, - "epoch": 4.0, + "epoch": 3.280224929709466, "eval_steps": 250, - "global_step": 8536, + "global_step": 7000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004686035613870665, - "grad_norm": 4.554049015045166, + "grad_norm": 0.5978041887283325, "learning_rate": 1.999707122774133e-05, - "loss": 2.7682, + "loss": 3.4332, "step": 10 }, { "epoch": 0.00937207122774133, - "grad_norm": 4.1577301025390625, + "grad_norm": 1.5130422115325928, "learning_rate": 1.9994142455482664e-05, - "loss": 2.0056, + "loss": 3.2454, "step": 20 }, { "epoch": 0.014058106841611996, - "grad_norm": 3.284484386444092, + "grad_norm": 1.607094168663025, "learning_rate": 1.9991213683223994e-05, - "loss": 1.4336, + "loss": 2.9102, "step": 30 }, { "epoch": 0.01874414245548266, - "grad_norm": 2.952692985534668, + "grad_norm": 2.6665945053100586, "learning_rate": 1.9988284910965326e-05, - "loss": 1.1235, + "loss": 2.7066, "step": 40 }, { "epoch": 0.023430178069353328, - "grad_norm": 2.8465845584869385, + "grad_norm": 3.212167978286743, "learning_rate": 1.9985356138706656e-05, - "loss": 1.014, + "loss": 2.5895, "step": 50 }, { "epoch": 0.028116213683223992, - "grad_norm": 2.830334424972534, + "grad_norm": 3.8364155292510986, "learning_rate": 1.998242736644799e-05, - "loss": 0.839, + "loss": 2.4215, "step": 60 }, { "epoch": 0.03280224929709466, - "grad_norm": 2.38661527633667, + "grad_norm": 3.560389280319214, "learning_rate": 1.9979498594189318e-05, - "loss": 0.7998, + "loss": 2.0972, "step": 70 }, { "epoch": 0.03748828491096532, - "grad_norm": 2.4744863510131836, + "grad_norm": 4.126266956329346, "learning_rate": 1.9976569821930648e-05, - "loss": 0.8126, + "loss": 1.8428, "step": 80 }, { "epoch": 0.04217432052483599, - "grad_norm": 2.545337200164795, + "grad_norm": 4.506312370300293, "learning_rate": 1.997364104967198e-05, - "loss": 0.7235, + "loss": 1.5853, "step": 90 }, { "epoch": 0.046860356138706656, - "grad_norm": 2.316910743713379, + "grad_norm": 4.499500751495361, "learning_rate": 1.997071227741331e-05, - "loss": 0.7556, + "loss": 1.4429, "step": 100 }, { "epoch": 0.05154639175257732, - "grad_norm": 2.520934581756592, + "grad_norm": 4.746804237365723, "learning_rate": 1.996778350515464e-05, - "loss": 0.687, + "loss": 1.3275, "step": 110 }, { "epoch": 0.056232427366447985, - "grad_norm": 2.330585479736328, + "grad_norm": 4.36592435836792, "learning_rate": 1.996485473289597e-05, - "loss": 0.5937, + "loss": 1.0965, "step": 120 }, { "epoch": 0.06091846298031865, - "grad_norm": 2.4207730293273926, + "grad_norm": 4.177589416503906, "learning_rate": 1.9961925960637302e-05, - "loss": 0.6238, + "loss": 1.0654, "step": 130 }, { "epoch": 0.06560449859418932, - "grad_norm": 2.1149699687957764, + "grad_norm": 4.377724647521973, "learning_rate": 1.995899718837863e-05, - "loss": 0.5714, + "loss": 0.9687, "step": 140 }, { "epoch": 0.07029053420805999, - "grad_norm": 2.2741851806640625, + "grad_norm": 3.9814295768737793, "learning_rate": 1.9956068416119964e-05, - "loss": 0.6148, + "loss": 1.043, "step": 150 }, { "epoch": 0.07497656982193064, - "grad_norm": 2.343580722808838, + "grad_norm": 3.938041925430298, "learning_rate": 1.9953139643861294e-05, - "loss": 0.5975, + "loss": 0.9639, "step": 160 }, { "epoch": 0.07966260543580131, - "grad_norm": 2.583735942840576, + "grad_norm": 4.323759078979492, "learning_rate": 1.9950210871602627e-05, - "loss": 0.5755, + "loss": 0.9344, "step": 170 }, { "epoch": 0.08434864104967198, - "grad_norm": 2.3597922325134277, + "grad_norm": 4.037902355194092, "learning_rate": 1.9947282099343956e-05, - "loss": 0.5139, + "loss": 0.8302, "step": 180 }, { "epoch": 0.08903467666354264, - "grad_norm": 2.2876176834106445, + "grad_norm": 3.9691343307495117, "learning_rate": 1.994435332708529e-05, - "loss": 0.5763, + "loss": 0.8573, "step": 190 }, { "epoch": 0.09372071227741331, - "grad_norm": 2.6187407970428467, + "grad_norm": 3.8372201919555664, "learning_rate": 1.994142455482662e-05, - "loss": 0.5075, + "loss": 0.808, "step": 200 }, { "epoch": 0.09840674789128398, - "grad_norm": 2.3355767726898193, + "grad_norm": 4.224605083465576, "learning_rate": 1.9938495782567948e-05, - "loss": 0.5719, + "loss": 0.8942, "step": 210 }, { "epoch": 0.10309278350515463, - "grad_norm": 1.9649430513381958, + "grad_norm": 3.7245891094207764, "learning_rate": 1.993556701030928e-05, - "loss": 0.5176, + "loss": 0.8321, "step": 220 }, { "epoch": 0.1077788191190253, - "grad_norm": 2.7376747131347656, + "grad_norm": 4.344135761260986, "learning_rate": 1.993263823805061e-05, - "loss": 0.5174, + "loss": 0.8035, "step": 230 }, { "epoch": 0.11246485473289597, - "grad_norm": 2.1346044540405273, + "grad_norm": 3.6320858001708984, "learning_rate": 1.9929709465791943e-05, - "loss": 0.4637, + "loss": 0.723, "step": 240 }, { "epoch": 0.11715089034676664, - "grad_norm": 1.9357292652130127, + "grad_norm": 2.934122085571289, "learning_rate": 1.9926780693533273e-05, - "loss": 0.4859, + "loss": 0.761, "step": 250 }, { "epoch": 0.11715089034676664, - "eval_loss": 0.07528682053089142, - "eval_pearson_cosine": 0.7922989999002419, - "eval_pearson_dot": 0.678508544067455, - "eval_pearson_euclidean": 0.7824953778206378, - "eval_pearson_manhattan": 0.78332634364466, - "eval_runtime": 3.139, - "eval_samples_per_second": 477.858, - "eval_spearman_cosine": 0.7923133107428318, - "eval_spearman_dot": 0.675670193412544, - "eval_spearman_euclidean": 0.7907281459858406, - "eval_spearman_manhattan": 0.7910650705518277, - "eval_steps_per_second": 29.946, + "eval_loss": 0.13966944813728333, + "eval_pearson_cosine": 0.7191493566389795, + "eval_pearson_dot": 0.43422507697410495, + "eval_pearson_euclidean": 0.7135360350736899, + "eval_pearson_manhattan": 0.7129270627309872, + "eval_runtime": 3.3956, + "eval_samples_per_second": 441.749, + "eval_spearman_cosine": 0.7365500082397669, + "eval_spearman_dot": 0.43017280395518565, + "eval_spearman_euclidean": 0.7209662281529704, + "eval_spearman_manhattan": 0.720539010106817, + "eval_steps_per_second": 27.683, "step": 250 }, { "epoch": 0.1218369259606373, - "grad_norm": 2.361877679824829, + "grad_norm": 3.1349332332611084, "learning_rate": 1.9923851921274602e-05, - "loss": 0.5312, + "loss": 0.7916, "step": 260 }, { "epoch": 0.12652296157450796, - "grad_norm": 2.8095247745513916, + "grad_norm": 4.0211286544799805, "learning_rate": 1.9920923149015935e-05, - "loss": 0.4791, + "loss": 0.7147, "step": 270 }, { "epoch": 0.13120899718837864, - "grad_norm": 2.2023308277130127, + "grad_norm": 3.262589693069458, "learning_rate": 1.9917994376757265e-05, - "loss": 0.4917, + "loss": 0.7689, "step": 280 }, { "epoch": 0.1358950328022493, - "grad_norm": 2.9753897190093994, + "grad_norm": 4.0925397872924805, "learning_rate": 1.9915065604498594e-05, - "loss": 0.505, + "loss": 0.7556, "step": 290 }, { "epoch": 0.14058106841611998, - "grad_norm": 2.363333225250244, + "grad_norm": 3.7041923999786377, "learning_rate": 1.9912136832239924e-05, - "loss": 0.4664, + "loss": 0.7239, "step": 300 }, { "epoch": 0.14526710402999063, - "grad_norm": 2.2978515625, + "grad_norm": 3.712703227996826, "learning_rate": 1.9909208059981257e-05, - "loss": 0.469, + "loss": 0.6685, "step": 310 }, { "epoch": 0.14995313964386128, - "grad_norm": 2.4117813110351562, + "grad_norm": 3.6240854263305664, "learning_rate": 1.9906279287722586e-05, - "loss": 0.4465, + "loss": 0.7263, "step": 320 }, { "epoch": 0.15463917525773196, - "grad_norm": 2.253214120864868, + "grad_norm": 4.013660907745361, "learning_rate": 1.990335051546392e-05, - "loss": 0.5049, + "loss": 0.7287, "step": 330 }, { "epoch": 0.15932521087160262, - "grad_norm": 2.005753755569458, + "grad_norm": 2.8780667781829834, "learning_rate": 1.990042174320525e-05, - "loss": 0.4656, + "loss": 0.6709, "step": 340 }, { "epoch": 0.1640112464854733, - "grad_norm": 1.8813114166259766, + "grad_norm": 3.035186290740967, "learning_rate": 1.989749297094658e-05, - "loss": 0.4848, + "loss": 0.6977, "step": 350 }, { "epoch": 0.16869728209934395, - "grad_norm": 2.469801187515259, + "grad_norm": 3.433403491973877, "learning_rate": 1.989456419868791e-05, - "loss": 0.4364, + "loss": 0.6656, "step": 360 }, { "epoch": 0.1733833177132146, - "grad_norm": 2.9116897583007812, + "grad_norm": 3.8357644081115723, "learning_rate": 1.9891635426429244e-05, - "loss": 0.4128, + "loss": 0.6192, "step": 370 }, { "epoch": 0.1780693533270853, - "grad_norm": 2.666409730911255, + "grad_norm": 3.506263494491577, "learning_rate": 1.9888706654170573e-05, - "loss": 0.4546, + "loss": 0.6516, "step": 380 }, { "epoch": 0.18275538894095594, - "grad_norm": 2.177461862564087, + "grad_norm": 2.839836835861206, "learning_rate": 1.9885777881911906e-05, - "loss": 0.441, + "loss": 0.6056, "step": 390 }, { "epoch": 0.18744142455482662, - "grad_norm": 2.3446977138519287, + "grad_norm": 3.2694239616394043, "learning_rate": 1.9882849109653236e-05, - "loss": 0.4446, + "loss": 0.6753, "step": 400 }, { "epoch": 0.19212746016869728, - "grad_norm": 3.112436056137085, + "grad_norm": 3.5731680393218994, "learning_rate": 1.9879920337394565e-05, - "loss": 0.4805, + "loss": 0.6848, "step": 410 }, { "epoch": 0.19681349578256796, - "grad_norm": 2.373749256134033, + "grad_norm": 2.8836023807525635, "learning_rate": 1.9876991565135898e-05, - "loss": 0.4271, + "loss": 0.659, "step": 420 }, { "epoch": 0.2014995313964386, - "grad_norm": 2.5556981563568115, + "grad_norm": 3.5604217052459717, "learning_rate": 1.9874062792877228e-05, - "loss": 0.4763, + "loss": 0.6745, "step": 430 }, { "epoch": 0.20618556701030927, - "grad_norm": 1.7662516832351685, + "grad_norm": 2.7834393978118896, "learning_rate": 1.9871134020618557e-05, - "loss": 0.4339, + "loss": 0.623, "step": 440 }, { "epoch": 0.21087160262417995, - "grad_norm": 2.5844173431396484, + "grad_norm": 3.5130224227905273, "learning_rate": 1.986820524835989e-05, - "loss": 0.4066, + "loss": 0.579, "step": 450 }, { "epoch": 0.2155576382380506, - "grad_norm": 2.0891411304473877, + "grad_norm": 2.9537723064422607, "learning_rate": 1.986527647610122e-05, - "loss": 0.439, + "loss": 0.6042, "step": 460 }, { "epoch": 0.22024367385192128, - "grad_norm": 1.9701979160308838, + "grad_norm": 2.926415205001831, "learning_rate": 1.986234770384255e-05, - "loss": 0.4249, + "loss": 0.5785, "step": 470 }, { "epoch": 0.22492970946579194, - "grad_norm": 2.371614456176758, + "grad_norm": 4.53303861618042, "learning_rate": 1.9859418931583882e-05, - "loss": 0.402, + "loss": 0.5957, "step": 480 }, { "epoch": 0.2296157450796626, - "grad_norm": 2.127275228500366, + "grad_norm": 2.5960335731506348, "learning_rate": 1.985649015932521e-05, - "loss": 0.4394, + "loss": 0.6223, "step": 490 }, { "epoch": 0.23430178069353327, - "grad_norm": 1.7096961736679077, + "grad_norm": 2.509232997894287, "learning_rate": 1.985356138706654e-05, - "loss": 0.4421, + "loss": 0.6275, "step": 500 }, { "epoch": 0.23430178069353327, - "eval_loss": 0.06987733393907547, - "eval_pearson_cosine": 0.7956408502961665, - "eval_pearson_dot": 0.6754197486337432, - "eval_pearson_euclidean": 0.7887092838983492, - "eval_pearson_manhattan": 0.78937785258519, - "eval_runtime": 3.5185, - "eval_samples_per_second": 426.316, - "eval_spearman_cosine": 0.7989362676746473, - "eval_spearman_dot": 0.6702224917590883, - "eval_spearman_euclidean": 0.7979524664607132, - "eval_spearman_manhattan": 0.7986639363570017, - "eval_steps_per_second": 26.716, + "eval_loss": 0.12402218580245972, + "eval_pearson_cosine": 0.7535085617332911, + "eval_pearson_dot": 0.4526570980382658, + "eval_pearson_euclidean": 0.7442016683051946, + "eval_pearson_manhattan": 0.7442320131099294, + "eval_runtime": 3.5327, + "eval_samples_per_second": 424.602, + "eval_spearman_cosine": 0.7638158285362802, + "eval_spearman_dot": 0.4533334110281352, + "eval_spearman_euclidean": 0.7505934110060137, + "eval_spearman_manhattan": 0.7505444222182343, + "eval_steps_per_second": 26.608, "step": 500 }, { "epoch": 0.23898781630740393, - "grad_norm": 2.0291731357574463, + "grad_norm": 3.2064850330352783, "learning_rate": 1.9850632614807874e-05, - "loss": 0.4377, + "loss": 0.6257, "step": 510 }, { "epoch": 0.2436738519212746, - "grad_norm": 2.478485345840454, + "grad_norm": 3.235630512237549, "learning_rate": 1.9847703842549203e-05, - "loss": 0.4347, + "loss": 0.6069, "step": 520 }, { "epoch": 0.24835988753514526, - "grad_norm": 1.9518004655838013, + "grad_norm": 2.8067846298217773, "learning_rate": 1.9844775070290536e-05, - "loss": 0.4334, + "loss": 0.6137, "step": 530 }, { "epoch": 0.2530459231490159, - "grad_norm": 2.35752272605896, + "grad_norm": 3.6333608627319336, "learning_rate": 1.9841846298031866e-05, - "loss": 0.4382, + "loss": 0.6111, "step": 540 }, { "epoch": 0.25773195876288657, - "grad_norm": 1.5922209024429321, + "grad_norm": 2.410837411880493, "learning_rate": 1.98389175257732e-05, - "loss": 0.3765, + "loss": 0.5463, "step": 550 }, { "epoch": 0.2624179943767573, - "grad_norm": 2.2074806690216064, + "grad_norm": 3.0617260932922363, "learning_rate": 1.9835988753514528e-05, - "loss": 0.4038, + "loss": 0.5571, "step": 560 }, { "epoch": 0.26710402999062793, - "grad_norm": 1.740545392036438, + "grad_norm": 2.576580762863159, "learning_rate": 1.983305998125586e-05, - "loss": 0.417, + "loss": 0.5428, "step": 570 }, { "epoch": 0.2717900656044986, - "grad_norm": 2.4346160888671875, + "grad_norm": 3.412797451019287, "learning_rate": 1.983013120899719e-05, - "loss": 0.3817, + "loss": 0.5541, "step": 580 }, { "epoch": 0.27647610121836924, - "grad_norm": 1.9497504234313965, + "grad_norm": 2.929755210876465, "learning_rate": 1.9827202436738523e-05, - "loss": 0.4312, + "loss": 0.5941, "step": 590 }, { "epoch": 0.28116213683223995, - "grad_norm": 2.0708413124084473, + "grad_norm": 2.80749773979187, "learning_rate": 1.9824273664479853e-05, - "loss": 0.4009, + "loss": 0.5702, "step": 600 }, { "epoch": 0.2858481724461106, - "grad_norm": 2.3257670402526855, + "grad_norm": 2.91896390914917, "learning_rate": 1.9821344892221182e-05, - "loss": 0.4131, + "loss": 0.5971, "step": 610 }, { "epoch": 0.29053420805998126, - "grad_norm": 1.5978585481643677, + "grad_norm": 2.754328727722168, "learning_rate": 1.9818416119962512e-05, - "loss": 0.3976, + "loss": 0.5544, "step": 620 }, { "epoch": 0.2952202436738519, - "grad_norm": 2.497614860534668, + "grad_norm": 3.4075613021850586, "learning_rate": 1.9815487347703845e-05, - "loss": 0.4166, + "loss": 0.5981, "step": 630 }, { "epoch": 0.29990627928772257, - "grad_norm": 2.02498459815979, + "grad_norm": 2.6344642639160156, "learning_rate": 1.9812558575445174e-05, - "loss": 0.4156, + "loss": 0.5539, "step": 640 }, { "epoch": 0.3045923149015933, - "grad_norm": 1.8266093730926514, + "grad_norm": 2.417914867401123, "learning_rate": 1.9809629803186504e-05, - "loss": 0.3699, + "loss": 0.5078, "step": 650 }, { "epoch": 0.30927835051546393, - "grad_norm": 1.82723069190979, + "grad_norm": 2.52506685256958, "learning_rate": 1.9806701030927837e-05, - "loss": 0.3677, + "loss": 0.532, "step": 660 }, { "epoch": 0.3139643861293346, - "grad_norm": 2.6589696407318115, + "grad_norm": 2.830923557281494, "learning_rate": 1.9803772258669166e-05, - "loss": 0.4123, + "loss": 0.5537, "step": 670 }, { "epoch": 0.31865042174320524, - "grad_norm": 1.9798301458358765, + "grad_norm": 3.1638379096984863, "learning_rate": 1.98008434864105e-05, - "loss": 0.3841, + "loss": 0.5375, "step": 680 }, { "epoch": 0.3233364573570759, - "grad_norm": 2.3881802558898926, + "grad_norm": 3.447112798690796, "learning_rate": 1.979791471415183e-05, - "loss": 0.4366, + "loss": 0.5755, "step": 690 }, { "epoch": 0.3280224929709466, - "grad_norm": 1.739403486251831, + "grad_norm": 2.3407933712005615, "learning_rate": 1.9794985941893158e-05, - "loss": 0.4062, + "loss": 0.514, "step": 700 }, { "epoch": 0.33270852858481725, - "grad_norm": 2.0091781616210938, + "grad_norm": 3.226551055908203, "learning_rate": 1.979205716963449e-05, - "loss": 0.3479, + "loss": 0.5167, "step": 710 }, { "epoch": 0.3373945641986879, - "grad_norm": 2.635582208633423, + "grad_norm": 3.1462602615356445, "learning_rate": 1.978912839737582e-05, - "loss": 0.403, + "loss": 0.5431, "step": 720 }, { "epoch": 0.34208059981255856, - "grad_norm": 1.9966530799865723, + "grad_norm": 3.051692008972168, "learning_rate": 1.9786199625117153e-05, - "loss": 0.3645, + "loss": 0.5129, "step": 730 }, { "epoch": 0.3467666354264292, - "grad_norm": 1.4709430932998657, + "grad_norm": 2.5202085971832275, "learning_rate": 1.9783270852858483e-05, - "loss": 0.3893, + "loss": 0.5752, "step": 740 }, { "epoch": 0.3514526710402999, - "grad_norm": 1.9793283939361572, + "grad_norm": 2.6891393661499023, "learning_rate": 1.9780342080599816e-05, - "loss": 0.3553, + "loss": 0.5326, "step": 750 }, { "epoch": 0.3514526710402999, - "eval_loss": 0.05558054894208908, - "eval_pearson_cosine": 0.8076353144789259, - "eval_pearson_dot": 0.7051323925422726, - "eval_pearson_euclidean": 0.802419363680495, - "eval_pearson_manhattan": 0.8035909791491171, - "eval_runtime": 3.8116, - "eval_samples_per_second": 393.533, - "eval_spearman_cosine": 0.8088217067659009, - "eval_spearman_dot": 0.7031203348796395, - "eval_spearman_euclidean": 0.809014079377687, - "eval_spearman_manhattan": 0.8095637114615523, - "eval_steps_per_second": 24.661, + "eval_loss": 0.11494184285402298, + "eval_pearson_cosine": 0.7539853173929743, + "eval_pearson_dot": 0.47863358738842976, + "eval_pearson_euclidean": 0.7326912597509363, + "eval_pearson_manhattan": 0.7320283236539638, + "eval_runtime": 3.5674, + "eval_samples_per_second": 420.478, + "eval_spearman_cosine": 0.7697587773995457, + "eval_spearman_dot": 0.47367550293168154, + "eval_spearman_euclidean": 0.7466430705786936, + "eval_spearman_manhattan": 0.7460682347775467, + "eval_steps_per_second": 26.35, "step": 750 }, { "epoch": 0.3561387066541706, - "grad_norm": 2.9553966522216797, + "grad_norm": 2.9754297733306885, "learning_rate": 1.9777413308341145e-05, - "loss": 0.3574, + "loss": 0.4959, "step": 760 }, { "epoch": 0.36082474226804123, - "grad_norm": 2.3012421131134033, + "grad_norm": 3.2427656650543213, "learning_rate": 1.9774484536082478e-05, - "loss": 0.4063, + "loss": 0.539, "step": 770 }, { "epoch": 0.3655107778819119, - "grad_norm": 1.9560497999191284, + "grad_norm": 2.6402335166931152, "learning_rate": 1.9771555763823808e-05, - "loss": 0.3636, + "loss": 0.5328, "step": 780 }, { "epoch": 0.3701968134957826, - "grad_norm": 1.8920570611953735, + "grad_norm": 2.9217915534973145, "learning_rate": 1.9768626991565137e-05, - "loss": 0.3283, + "loss": 0.4895, "step": 790 }, { "epoch": 0.37488284910965325, - "grad_norm": 2.03069806098938, + "grad_norm": 2.363999128341675, "learning_rate": 1.976569821930647e-05, - "loss": 0.3643, + "loss": 0.4805, "step": 800 }, { "epoch": 0.3795688847235239, - "grad_norm": 1.8706809282302856, + "grad_norm": 2.491255044937134, "learning_rate": 1.97627694470478e-05, - "loss": 0.4174, + "loss": 0.5281, "step": 810 }, { "epoch": 0.38425492033739456, - "grad_norm": 1.7476972341537476, + "grad_norm": 2.595560312271118, "learning_rate": 1.975984067478913e-05, - "loss": 0.3594, + "loss": 0.5104, "step": 820 }, { "epoch": 0.3889409559512652, - "grad_norm": 1.849311351776123, + "grad_norm": 2.991326332092285, "learning_rate": 1.975691190253046e-05, - "loss": 0.3581, + "loss": 0.508, "step": 830 }, { "epoch": 0.3936269915651359, - "grad_norm": 2.033726692199707, + "grad_norm": 2.6440255641937256, "learning_rate": 1.975398313027179e-05, - "loss": 0.373, + "loss": 0.5643, "step": 840 }, { "epoch": 0.3983130271790066, - "grad_norm": 1.9408748149871826, + "grad_norm": 3.06538987159729, "learning_rate": 1.975105435801312e-05, - "loss": 0.34, + "loss": 0.4944, "step": 850 }, { "epoch": 0.4029990627928772, - "grad_norm": 1.997316837310791, + "grad_norm": 2.751331329345703, "learning_rate": 1.9748125585754454e-05, - "loss": 0.3757, + "loss": 0.5215, "step": 860 }, { "epoch": 0.4076850984067479, - "grad_norm": 2.345407247543335, + "grad_norm": 3.2387125492095947, "learning_rate": 1.9745196813495783e-05, - "loss": 0.4194, + "loss": 0.5925, "step": 870 }, { "epoch": 0.41237113402061853, - "grad_norm": 2.0166800022125244, + "grad_norm": 2.5455057621002197, "learning_rate": 1.9742268041237116e-05, - "loss": 0.3778, + "loss": 0.4956, "step": 880 }, { "epoch": 0.41705716963448924, - "grad_norm": 2.222252130508423, + "grad_norm": 2.864497423171997, "learning_rate": 1.9739339268978446e-05, - "loss": 0.3678, + "loss": 0.4855, "step": 890 }, { "epoch": 0.4217432052483599, - "grad_norm": 1.6460472345352173, + "grad_norm": 2.6306967735290527, "learning_rate": 1.9736410496719775e-05, - "loss": 0.346, + "loss": 0.4542, "step": 900 }, { "epoch": 0.42642924086223055, - "grad_norm": 1.80186927318573, + "grad_norm": 2.294090747833252, "learning_rate": 1.9733481724461108e-05, - "loss": 0.2925, + "loss": 0.4247, "step": 910 }, { "epoch": 0.4311152764761012, - "grad_norm": 2.294146776199341, + "grad_norm": 2.9401843547821045, "learning_rate": 1.9730552952202438e-05, - "loss": 0.3908, + "loss": 0.5219, "step": 920 }, { "epoch": 0.43580131208997186, - "grad_norm": 2.041172742843628, + "grad_norm": 2.5331194400787354, "learning_rate": 1.972762417994377e-05, - "loss": 0.3616, + "loss": 0.5193, "step": 930 }, { "epoch": 0.44048734770384257, - "grad_norm": 1.9510180950164795, + "grad_norm": 3.265944719314575, "learning_rate": 1.97246954076851e-05, - "loss": 0.3135, + "loss": 0.4733, "step": 940 }, { "epoch": 0.4451733833177132, - "grad_norm": 1.9195556640625, + "grad_norm": 2.6586129665374756, "learning_rate": 1.9721766635426433e-05, - "loss": 0.3555, + "loss": 0.4939, "step": 950 }, { "epoch": 0.4498594189315839, - "grad_norm": 1.6835869550704956, + "grad_norm": 2.126664638519287, "learning_rate": 1.9718837863167762e-05, - "loss": 0.3769, + "loss": 0.5149, "step": 960 }, { "epoch": 0.45454545454545453, - "grad_norm": 2.1964399814605713, + "grad_norm": 2.8597347736358643, "learning_rate": 1.9715909090909092e-05, - "loss": 0.3758, + "loss": 0.5089, "step": 970 }, { "epoch": 0.4592314901593252, - "grad_norm": 1.761430263519287, + "grad_norm": 2.5802338123321533, "learning_rate": 1.9712980318650425e-05, - "loss": 0.3738, + "loss": 0.5264, "step": 980 }, { "epoch": 0.4639175257731959, - "grad_norm": 1.6870702505111694, + "grad_norm": 2.4247639179229736, "learning_rate": 1.9710051546391754e-05, - "loss": 0.3873, + "loss": 0.5416, "step": 990 }, { "epoch": 0.46860356138706655, - "grad_norm": 1.9529205560684204, + "grad_norm": 2.6614298820495605, "learning_rate": 1.9707122774133084e-05, - "loss": 0.3311, + "loss": 0.4917, "step": 1000 }, { "epoch": 0.46860356138706655, - "eval_loss": 0.05577142909169197, - "eval_pearson_cosine": 0.8113553088294765, - "eval_pearson_dot": 0.7185301210033685, - "eval_pearson_euclidean": 0.8040205194581489, - "eval_pearson_manhattan": 0.8050175035364262, - "eval_runtime": 4.3673, - "eval_samples_per_second": 343.461, - "eval_spearman_cosine": 0.8143275202049945, - "eval_spearman_dot": 0.7185323988486403, - "eval_spearman_euclidean": 0.8117920593159309, - "eval_spearman_manhattan": 0.8126305556244356, - "eval_steps_per_second": 21.524, + "eval_loss": 0.1028011292219162, + "eval_pearson_cosine": 0.7630335129215418, + "eval_pearson_dot": 0.5427640329363896, + "eval_pearson_euclidean": 0.7394698860945113, + "eval_pearson_manhattan": 0.7395035409540256, + "eval_runtime": 3.2727, + "eval_samples_per_second": 458.333, + "eval_spearman_cosine": 0.7777749785163052, + "eval_spearman_dot": 0.5403903388556195, + "eval_spearman_euclidean": 0.7530551416243708, + "eval_spearman_manhattan": 0.753178994867466, + "eval_steps_per_second": 28.722, "step": 1000 }, { "epoch": 0.4732895970009372, - "grad_norm": 2.142782211303711, + "grad_norm": 2.8377857208251953, "learning_rate": 1.9704194001874417e-05, - "loss": 0.3473, + "loss": 0.4806, "step": 1010 }, { "epoch": 0.47797563261480785, - "grad_norm": 1.7832313776016235, + "grad_norm": 2.3730361461639404, "learning_rate": 1.9701265229615746e-05, - "loss": 0.3601, + "loss": 0.4946, "step": 1020 }, { "epoch": 0.48266166822867856, - "grad_norm": 1.9616563320159912, + "grad_norm": 3.101628541946411, "learning_rate": 1.9698336457357076e-05, - "loss": 0.3489, + "loss": 0.4958, "step": 1030 }, { "epoch": 0.4873477038425492, - "grad_norm": 2.368060827255249, + "grad_norm": 2.4200384616851807, "learning_rate": 1.969540768509841e-05, - "loss": 0.3955, + "loss": 0.5231, "step": 1040 }, { "epoch": 0.49203373945641987, - "grad_norm": 2.006263017654419, + "grad_norm": 2.9535088539123535, "learning_rate": 1.9692478912839738e-05, - "loss": 0.3339, + "loss": 0.4952, "step": 1050 }, { "epoch": 0.4967197750702905, - "grad_norm": 2.0187647342681885, + "grad_norm": 2.3175110816955566, "learning_rate": 1.968955014058107e-05, - "loss": 0.3491, + "loss": 0.4845, "step": 1060 }, { "epoch": 0.5014058106841612, - "grad_norm": 1.87997305393219, + "grad_norm": 2.934502124786377, "learning_rate": 1.96866213683224e-05, - "loss": 0.3784, + "loss": 0.5229, "step": 1070 }, { "epoch": 0.5060918462980318, - "grad_norm": 2.165182113647461, + "grad_norm": 2.5961878299713135, "learning_rate": 1.9683692596063733e-05, - "loss": 0.3784, + "loss": 0.5111, "step": 1080 }, { "epoch": 0.5107778819119025, - "grad_norm": 2.0993471145629883, + "grad_norm": 2.9093525409698486, "learning_rate": 1.9680763823805063e-05, - "loss": 0.3719, + "loss": 0.545, "step": 1090 }, { "epoch": 0.5154639175257731, - "grad_norm": 2.335336923599243, + "grad_norm": 2.733900547027588, "learning_rate": 1.9677835051546392e-05, - "loss": 0.3357, + "loss": 0.4589, "step": 1100 }, { "epoch": 0.5201499531396439, - "grad_norm": 2.261863946914673, + "grad_norm": 2.3905482292175293, "learning_rate": 1.9674906279287725e-05, - "loss": 0.3678, + "loss": 0.5197, "step": 1110 }, { "epoch": 0.5248359887535146, - "grad_norm": 1.9904841184616089, + "grad_norm": 2.5557029247283936, "learning_rate": 1.9671977507029055e-05, - "loss": 0.3646, + "loss": 0.5089, "step": 1120 }, { "epoch": 0.5295220243673852, - "grad_norm": 1.7399210929870605, + "grad_norm": 2.3789846897125244, "learning_rate": 1.9669048734770388e-05, - "loss": 0.3927, + "loss": 0.5373, "step": 1130 }, { "epoch": 0.5342080599812559, - "grad_norm": 2.1144919395446777, + "grad_norm": 2.848998785018921, "learning_rate": 1.9666119962511717e-05, - "loss": 0.4183, + "loss": 0.5321, "step": 1140 }, { "epoch": 0.5388940955951266, - "grad_norm": 1.5780616998672485, + "grad_norm": 2.2063546180725098, "learning_rate": 1.9663191190253047e-05, - "loss": 0.3507, + "loss": 0.4809, "step": 1150 }, { "epoch": 0.5435801312089972, - "grad_norm": 1.9400485754013062, + "grad_norm": 2.7333438396453857, "learning_rate": 1.966026241799438e-05, - "loss": 0.3326, + "loss": 0.4615, "step": 1160 }, { "epoch": 0.5482661668228679, - "grad_norm": 2.7908225059509277, + "grad_norm": 3.124687671661377, "learning_rate": 1.965733364573571e-05, - "loss": 0.354, + "loss": 0.5177, "step": 1170 }, { "epoch": 0.5529522024367385, - "grad_norm": 1.672563910484314, + "grad_norm": 2.726681709289551, "learning_rate": 1.965440487347704e-05, - "loss": 0.3708, + "loss": 0.4891, "step": 1180 }, { "epoch": 0.5576382380506092, - "grad_norm": 2.061281681060791, + "grad_norm": 2.9802463054656982, "learning_rate": 1.965147610121837e-05, - "loss": 0.3445, + "loss": 0.5018, "step": 1190 }, { "epoch": 0.5623242736644799, - "grad_norm": 2.315765857696533, + "grad_norm": 2.844006299972534, "learning_rate": 1.96485473289597e-05, - "loss": 0.376, + "loss": 0.5149, "step": 1200 }, { "epoch": 0.5670103092783505, - "grad_norm": 2.0039291381835938, + "grad_norm": 2.442472219467163, "learning_rate": 1.964561855670103e-05, - "loss": 0.333, + "loss": 0.4808, "step": 1210 }, { "epoch": 0.5716963448922212, - "grad_norm": 1.721211552619934, + "grad_norm": 1.9972388744354248, "learning_rate": 1.9642689784442363e-05, - "loss": 0.344, + "loss": 0.4743, "step": 1220 }, { "epoch": 0.5763823805060918, - "grad_norm": 1.9561458826065063, + "grad_norm": 2.6430776119232178, "learning_rate": 1.9639761012183693e-05, - "loss": 0.3472, + "loss": 0.5057, "step": 1230 }, { "epoch": 0.5810684161199625, - "grad_norm": 1.6850024461746216, + "grad_norm": 2.37652850151062, "learning_rate": 1.9636832239925026e-05, - "loss": 0.3101, + "loss": 0.4851, "step": 1240 }, { "epoch": 0.5857544517338332, - "grad_norm": 2.0207340717315674, + "grad_norm": 2.755315065383911, "learning_rate": 1.9633903467666355e-05, - "loss": 0.3541, + "loss": 0.4451, "step": 1250 }, { "epoch": 0.5857544517338332, - "eval_loss": 0.055569760501384735, - "eval_pearson_cosine": 0.806980140855714, - "eval_pearson_dot": 0.7040136144947127, - "eval_pearson_euclidean": 0.8125674124413185, - "eval_pearson_manhattan": 0.8134596639320257, - "eval_runtime": 4.5743, - "eval_samples_per_second": 327.922, - "eval_spearman_cosine": 0.8099314919725019, - "eval_spearman_dot": 0.701768085329828, - "eval_spearman_euclidean": 0.8179511234249097, - "eval_spearman_manhattan": 0.8183207328526064, - "eval_steps_per_second": 20.55, + "eval_loss": 0.09590280055999756, + "eval_pearson_cosine": 0.7634019154373535, + "eval_pearson_dot": 0.5909367397822605, + "eval_pearson_euclidean": 0.7507597414427352, + "eval_pearson_manhattan": 0.7505103943331619, + "eval_runtime": 3.3916, + "eval_samples_per_second": 442.267, + "eval_spearman_cosine": 0.7803407693262222, + "eval_spearman_dot": 0.592884901593322, + "eval_spearman_euclidean": 0.7652017734526011, + "eval_spearman_manhattan": 0.7648713625602643, + "eval_steps_per_second": 27.715, "step": 1250 }, { "epoch": 0.5904404873477038, - "grad_norm": 2.3315682411193848, + "grad_norm": 3.0767931938171387, "learning_rate": 1.9630974695407688e-05, - "loss": 0.3762, + "loss": 0.5093, "step": 1260 }, { "epoch": 0.5951265229615745, - "grad_norm": 2.271893262863159, + "grad_norm": 3.0984067916870117, "learning_rate": 1.9628045923149017e-05, - "loss": 0.372, + "loss": 0.4888, "step": 1270 }, { "epoch": 0.5998125585754451, - "grad_norm": 1.830249309539795, + "grad_norm": 2.496957778930664, "learning_rate": 1.962511715089035e-05, - "loss": 0.3228, + "loss": 0.5005, "step": 1280 }, { "epoch": 0.6044985941893158, - "grad_norm": 1.4407238960266113, + "grad_norm": 2.1109137535095215, "learning_rate": 1.962218837863168e-05, - "loss": 0.3235, + "loss": 0.4729, "step": 1290 }, { "epoch": 0.6091846298031866, - "grad_norm": 2.0112946033477783, + "grad_norm": 2.663804054260254, "learning_rate": 1.961925960637301e-05, - "loss": 0.3107, + "loss": 0.4454, "step": 1300 }, { "epoch": 0.6138706654170571, - "grad_norm": 1.486770749092102, + "grad_norm": 1.8708107471466064, "learning_rate": 1.9616330834114342e-05, - "loss": 0.3695, + "loss": 0.4949, "step": 1310 }, { "epoch": 0.6185567010309279, - "grad_norm": 2.227010726928711, + "grad_norm": 2.9178926944732666, "learning_rate": 1.9613402061855672e-05, - "loss": 0.3301, + "loss": 0.4438, "step": 1320 }, { "epoch": 0.6232427366447985, - "grad_norm": 1.4869340658187866, + "grad_norm": 2.1928579807281494, "learning_rate": 1.9610473289597005e-05, - "loss": 0.3289, + "loss": 0.4483, "step": 1330 }, { "epoch": 0.6279287722586692, - "grad_norm": 1.6876165866851807, + "grad_norm": 2.3122403621673584, "learning_rate": 1.9607544517338334e-05, - "loss": 0.3445, + "loss": 0.4677, "step": 1340 }, { "epoch": 0.6326148078725399, - "grad_norm": 2.896570920944214, + "grad_norm": 2.5242831707000732, "learning_rate": 1.9604615745079664e-05, - "loss": 0.354, + "loss": 0.4822, "step": 1350 }, { "epoch": 0.6373008434864105, - "grad_norm": 1.7586398124694824, + "grad_norm": 2.572446346282959, "learning_rate": 1.9601686972820993e-05, - "loss": 0.305, + "loss": 0.46, "step": 1360 }, { "epoch": 0.6419868791002812, - "grad_norm": 1.5839918851852417, + "grad_norm": 2.6114673614501953, "learning_rate": 1.9598758200562326e-05, - "loss": 0.3382, + "loss": 0.4606, "step": 1370 }, { "epoch": 0.6466729147141518, - "grad_norm": 2.0355846881866455, + "grad_norm": 2.6151556968688965, "learning_rate": 1.9595829428303655e-05, - "loss": 0.3301, + "loss": 0.459, "step": 1380 }, { "epoch": 0.6513589503280225, - "grad_norm": 2.261348009109497, + "grad_norm": 2.9040744304656982, "learning_rate": 1.959290065604499e-05, - "loss": 0.355, + "loss": 0.4994, "step": 1390 }, { "epoch": 0.6560449859418932, - "grad_norm": 2.7674410343170166, + "grad_norm": 2.9549591541290283, "learning_rate": 1.9589971883786318e-05, - "loss": 0.2998, + "loss": 0.4335, "step": 1400 }, { "epoch": 0.6607310215557638, - "grad_norm": 1.7338224649429321, + "grad_norm": 2.2790753841400146, "learning_rate": 1.9587043111527647e-05, - "loss": 0.3323, + "loss": 0.4642, "step": 1410 }, { "epoch": 0.6654170571696345, - "grad_norm": 2.014716148376465, + "grad_norm": 2.265308141708374, "learning_rate": 1.958411433926898e-05, - "loss": 0.3342, + "loss": 0.4654, "step": 1420 }, { "epoch": 0.6701030927835051, - "grad_norm": 1.8897966146469116, + "grad_norm": 2.1528425216674805, "learning_rate": 1.958118556701031e-05, - "loss": 0.3592, + "loss": 0.4983, "step": 1430 }, { "epoch": 0.6747891283973758, - "grad_norm": 2.0015108585357666, + "grad_norm": 2.840533494949341, "learning_rate": 1.9578256794751643e-05, - "loss": 0.3577, + "loss": 0.4989, "step": 1440 }, { "epoch": 0.6794751640112465, - "grad_norm": 1.5381790399551392, + "grad_norm": 2.4118032455444336, "learning_rate": 1.9575328022492972e-05, - "loss": 0.3087, + "loss": 0.4207, "step": 1450 }, { "epoch": 0.6841611996251171, - "grad_norm": 1.7423696517944336, + "grad_norm": 2.845712900161743, "learning_rate": 1.9572399250234305e-05, - "loss": 0.3111, + "loss": 0.4219, "step": 1460 }, { "epoch": 0.6888472352389878, - "grad_norm": 1.9198020696640015, + "grad_norm": 2.6068544387817383, "learning_rate": 1.9569470477975635e-05, - "loss": 0.3442, + "loss": 0.4618, "step": 1470 }, { "epoch": 0.6935332708528584, - "grad_norm": 2.286470413208008, + "grad_norm": 2.6239466667175293, "learning_rate": 1.9566541705716967e-05, - "loss": 0.3347, + "loss": 0.4613, "step": 1480 }, { "epoch": 0.6982193064667291, - "grad_norm": 1.8658415079116821, + "grad_norm": 2.417541742324829, "learning_rate": 1.9563612933458297e-05, - "loss": 0.3026, + "loss": 0.4457, "step": 1490 }, { "epoch": 0.7029053420805998, - "grad_norm": 1.7906545400619507, + "grad_norm": 2.444243907928467, "learning_rate": 1.9560684161199626e-05, - "loss": 0.344, + "loss": 0.4682, "step": 1500 }, { "epoch": 0.7029053420805998, - "eval_loss": 0.054905861616134644, - "eval_pearson_cosine": 0.8153011005303412, - "eval_pearson_dot": 0.7054042470638251, - "eval_pearson_euclidean": 0.8096768473825762, - "eval_pearson_manhattan": 0.8108787022859651, - "eval_runtime": 5.1821, - "eval_samples_per_second": 289.46, - "eval_spearman_cosine": 0.8197251939965106, - "eval_spearman_dot": 0.7077519448371881, - "eval_spearman_euclidean": 0.8188198441060672, - "eval_spearman_manhattan": 0.8201609157379989, - "eval_steps_per_second": 18.139, + "eval_loss": 0.10573244839906693, + "eval_pearson_cosine": 0.7687240391174335, + "eval_pearson_dot": 0.52706539661569, + "eval_pearson_euclidean": 0.7544794171981408, + "eval_pearson_manhattan": 0.7540749762273435, + "eval_runtime": 3.5044, + "eval_samples_per_second": 428.028, + "eval_spearman_cosine": 0.7854868415020917, + "eval_spearman_dot": 0.5189687789236848, + "eval_spearman_euclidean": 0.7685180176151998, + "eval_spearman_manhattan": 0.7680518081256867, + "eval_steps_per_second": 26.823, "step": 1500 }, { "epoch": 0.7075913776944704, - "grad_norm": 1.9141615629196167, + "grad_norm": 2.4107391834259033, "learning_rate": 1.955775538894096e-05, - "loss": 0.2948, + "loss": 0.4231, "step": 1510 }, { "epoch": 0.7122774133083412, - "grad_norm": 1.94936203956604, + "grad_norm": 2.2897465229034424, "learning_rate": 1.955482661668229e-05, - "loss": 0.3326, + "loss": 0.495, "step": 1520 }, { "epoch": 0.7169634489222118, - "grad_norm": 1.8304425477981567, + "grad_norm": 2.1320059299468994, "learning_rate": 1.9551897844423618e-05, - "loss": 0.342, + "loss": 0.4362, "step": 1530 }, { "epoch": 0.7216494845360825, - "grad_norm": 1.7926722764968872, + "grad_norm": 2.1766531467437744, "learning_rate": 1.9548969072164948e-05, - "loss": 0.3213, + "loss": 0.4324, "step": 1540 }, { "epoch": 0.7263355201499532, - "grad_norm": 2.0607831478118896, + "grad_norm": 2.1343445777893066, "learning_rate": 1.954604029990628e-05, - "loss": 0.3136, + "loss": 0.4356, "step": 1550 }, { "epoch": 0.7310215557638238, - "grad_norm": 2.2151925563812256, + "grad_norm": 2.6357688903808594, "learning_rate": 1.954311152764761e-05, - "loss": 0.3415, + "loss": 0.4617, "step": 1560 }, { "epoch": 0.7357075913776945, - "grad_norm": 1.7518079280853271, + "grad_norm": 2.2863566875457764, "learning_rate": 1.9540182755388943e-05, - "loss": 0.3286, + "loss": 0.4444, "step": 1570 }, { "epoch": 0.7403936269915652, - "grad_norm": 1.475696086883545, + "grad_norm": 1.9845598936080933, "learning_rate": 1.9537253983130273e-05, - "loss": 0.3085, + "loss": 0.4459, "step": 1580 }, { "epoch": 0.7450796626054358, - "grad_norm": 1.8179190158843994, + "grad_norm": 2.3808443546295166, "learning_rate": 1.9534325210871605e-05, - "loss": 0.3498, + "loss": 0.4664, "step": 1590 }, { "epoch": 0.7497656982193065, - "grad_norm": 2.115126609802246, + "grad_norm": 2.976896286010742, "learning_rate": 1.9531396438612935e-05, - "loss": 0.3459, + "loss": 0.4317, "step": 1600 }, { "epoch": 0.7544517338331771, - "grad_norm": 2.411120891571045, + "grad_norm": 3.114450693130493, "learning_rate": 1.9528467666354264e-05, - "loss": 0.3146, + "loss": 0.4365, "step": 1610 }, { "epoch": 0.7591377694470478, - "grad_norm": 1.7418440580368042, + "grad_norm": 2.208798408508301, "learning_rate": 1.9525538894095597e-05, - "loss": 0.3269, + "loss": 0.4357, "step": 1620 }, { "epoch": 0.7638238050609185, - "grad_norm": 1.7588366270065308, + "grad_norm": 3.005563259124756, "learning_rate": 1.9522610121836927e-05, - "loss": 0.308, + "loss": 0.4184, "step": 1630 }, { "epoch": 0.7685098406747891, - "grad_norm": 1.7938318252563477, + "grad_norm": 2.711338758468628, "learning_rate": 1.951968134957826e-05, - "loss": 0.3261, + "loss": 0.4479, "step": 1640 }, { "epoch": 0.7731958762886598, - "grad_norm": 1.5655628442764282, + "grad_norm": 1.9399685859680176, "learning_rate": 1.951675257731959e-05, - "loss": 0.3519, + "loss": 0.4642, "step": 1650 }, { "epoch": 0.7778819119025304, - "grad_norm": 1.9560043811798096, + "grad_norm": 2.5553548336029053, "learning_rate": 1.9513823805060922e-05, - "loss": 0.3808, + "loss": 0.4851, "step": 1660 }, { "epoch": 0.7825679475164011, - "grad_norm": 2.287410259246826, + "grad_norm": 2.7015600204467773, "learning_rate": 1.951089503280225e-05, - "loss": 0.3063, + "loss": 0.4294, "step": 1670 }, { "epoch": 0.7872539831302718, - "grad_norm": 2.3901917934417725, + "grad_norm": 3.027247428894043, "learning_rate": 1.950796626054358e-05, - "loss": 0.3471, + "loss": 0.4359, "step": 1680 }, { "epoch": 0.7919400187441424, - "grad_norm": 2.0776314735412598, + "grad_norm": 2.603200674057007, "learning_rate": 1.9505037488284914e-05, - "loss": 0.3487, + "loss": 0.4833, "step": 1690 }, { "epoch": 0.7966260543580131, - "grad_norm": 1.6857753992080688, + "grad_norm": 2.2245004177093506, "learning_rate": 1.9502108716026243e-05, - "loss": 0.3546, + "loss": 0.4633, "step": 1700 }, { "epoch": 0.8013120899718837, - "grad_norm": 1.8472001552581787, + "grad_norm": 2.1866180896759033, "learning_rate": 1.9499179943767573e-05, - "loss": 0.3383, + "loss": 0.4465, "step": 1710 }, { "epoch": 0.8059981255857545, - "grad_norm": 2.173574447631836, + "grad_norm": 2.464083671569824, "learning_rate": 1.9496251171508906e-05, - "loss": 0.3201, + "loss": 0.4282, "step": 1720 }, { "epoch": 0.8106841611996252, - "grad_norm": 2.0859599113464355, + "grad_norm": 2.5016162395477295, "learning_rate": 1.9493322399250235e-05, - "loss": 0.3453, + "loss": 0.4566, "step": 1730 }, { "epoch": 0.8153701968134958, - "grad_norm": 1.829081416130066, + "grad_norm": 2.9877169132232666, "learning_rate": 1.9490393626991565e-05, - "loss": 0.33, + "loss": 0.462, "step": 1740 }, { "epoch": 0.8200562324273665, - "grad_norm": 1.9932607412338257, + "grad_norm": 2.2195188999176025, "learning_rate": 1.9487464854732898e-05, - "loss": 0.3268, + "loss": 0.4489, "step": 1750 }, { "epoch": 0.8200562324273665, - "eval_loss": 0.05351157858967781, - "eval_pearson_cosine": 0.817182487447738, - "eval_pearson_dot": 0.7224077847006214, - "eval_pearson_euclidean": 0.8127968942100594, - "eval_pearson_manhattan": 0.8137800174892043, - "eval_runtime": 4.5676, - "eval_samples_per_second": 328.4, - "eval_spearman_cosine": 0.8209700621713334, - "eval_spearman_dot": 0.7208139231824556, - "eval_spearman_euclidean": 0.8201513074791367, - "eval_spearman_manhattan": 0.8211217573317043, - "eval_steps_per_second": 20.58, + "eval_loss": 0.09936515241861343, + "eval_pearson_cosine": 0.7657869832789288, + "eval_pearson_dot": 0.576463122226734, + "eval_pearson_euclidean": 0.7513952467595857, + "eval_pearson_manhattan": 0.7505485236956133, + "eval_runtime": 3.3329, + "eval_samples_per_second": 450.052, + "eval_spearman_cosine": 0.7800032031654975, + "eval_spearman_dot": 0.576004160544721, + "eval_spearman_euclidean": 0.7627214415950069, + "eval_spearman_manhattan": 0.7623633881575729, + "eval_steps_per_second": 28.203, "step": 1750 }, { "epoch": 0.8247422680412371, - "grad_norm": 2.0316638946533203, + "grad_norm": 2.793635606765747, "learning_rate": 1.9484536082474227e-05, - "loss": 0.3236, + "loss": 0.4434, "step": 1760 }, { "epoch": 0.8294283036551078, - "grad_norm": 1.921434760093689, + "grad_norm": 2.403629779815674, "learning_rate": 1.948160731021556e-05, - "loss": 0.2945, + "loss": 0.4034, "step": 1770 }, { "epoch": 0.8341143392689785, - "grad_norm": 2.049253463745117, + "grad_norm": 2.6248741149902344, "learning_rate": 1.947867853795689e-05, - "loss": 0.3147, + "loss": 0.4328, "step": 1780 }, { "epoch": 0.8388003748828491, - "grad_norm": 1.5736606121063232, + "grad_norm": 1.8423967361450195, "learning_rate": 1.9475749765698223e-05, - "loss": 0.3209, + "loss": 0.458, "step": 1790 }, { "epoch": 0.8434864104967198, - "grad_norm": 1.7810872793197632, + "grad_norm": 2.8771181106567383, "learning_rate": 1.9472820993439552e-05, - "loss": 0.2527, + "loss": 0.3796, "step": 1800 }, { "epoch": 0.8481724461105904, - "grad_norm": 2.1259636878967285, + "grad_norm": 2.6302261352539062, "learning_rate": 1.946989222118088e-05, - "loss": 0.3173, + "loss": 0.4494, "step": 1810 }, { "epoch": 0.8528584817244611, - "grad_norm": 1.7195252180099487, + "grad_norm": 1.8728386163711548, "learning_rate": 1.9466963448922214e-05, - "loss": 0.3231, + "loss": 0.4624, "step": 1820 }, { "epoch": 0.8575445173383318, - "grad_norm": 2.140366792678833, + "grad_norm": 1.9615451097488403, "learning_rate": 1.9464034676663544e-05, - "loss": 0.3139, + "loss": 0.4666, "step": 1830 }, { "epoch": 0.8622305529522024, - "grad_norm": 2.2039151191711426, + "grad_norm": 2.423076629638672, "learning_rate": 1.9461105904404877e-05, - "loss": 0.3329, + "loss": 0.4659, "step": 1840 }, { "epoch": 0.8669165885660731, - "grad_norm": 1.644698143005371, + "grad_norm": 2.1707210540771484, "learning_rate": 1.9458177132146206e-05, - "loss": 0.3214, + "loss": 0.4536, "step": 1850 }, { "epoch": 0.8716026241799437, - "grad_norm": 2.189842939376831, + "grad_norm": 2.3241922855377197, "learning_rate": 1.9455248359887536e-05, - "loss": 0.3373, + "loss": 0.4589, "step": 1860 }, { "epoch": 0.8762886597938144, - "grad_norm": 1.977828025817871, + "grad_norm": 2.628664255142212, "learning_rate": 1.945231958762887e-05, - "loss": 0.3131, + "loss": 0.4801, "step": 1870 }, { "epoch": 0.8809746954076851, - "grad_norm": 1.8575206995010376, + "grad_norm": 2.25126314163208, "learning_rate": 1.9449390815370198e-05, - "loss": 0.3125, + "loss": 0.4127, "step": 1880 }, { "epoch": 0.8856607310215557, - "grad_norm": 2.189164638519287, + "grad_norm": 2.988956928253174, "learning_rate": 1.9446462043111528e-05, - "loss": 0.3328, + "loss": 0.4431, "step": 1890 }, { "epoch": 0.8903467666354264, - "grad_norm": 1.9307777881622314, + "grad_norm": 2.3071420192718506, "learning_rate": 1.944353327085286e-05, - "loss": 0.336, + "loss": 0.4606, "step": 1900 }, { "epoch": 0.895032802249297, - "grad_norm": 2.635529041290283, + "grad_norm": 2.712460517883301, "learning_rate": 1.944060449859419e-05, - "loss": 0.3469, + "loss": 0.4853, "step": 1910 }, { "epoch": 0.8997188378631678, - "grad_norm": 2.0925216674804688, + "grad_norm": 2.555626153945923, "learning_rate": 1.943767572633552e-05, - "loss": 0.3194, + "loss": 0.4187, "step": 1920 }, { "epoch": 0.9044048734770385, - "grad_norm": 1.815006136894226, + "grad_norm": 2.3509678840637207, "learning_rate": 1.9434746954076852e-05, - "loss": 0.3124, + "loss": 0.4144, "step": 1930 }, { "epoch": 0.9090909090909091, - "grad_norm": 2.069241762161255, + "grad_norm": 2.63004994392395, "learning_rate": 1.9431818181818182e-05, - "loss": 0.2978, + "loss": 0.399, "step": 1940 }, { "epoch": 0.9137769447047798, - "grad_norm": 1.9098782539367676, + "grad_norm": 2.8148534297943115, "learning_rate": 1.9428889409559515e-05, - "loss": 0.298, + "loss": 0.4309, "step": 1950 }, { "epoch": 0.9184629803186504, - "grad_norm": 2.433910369873047, + "grad_norm": 2.594778537750244, "learning_rate": 1.9425960637300844e-05, - "loss": 0.3403, + "loss": 0.4753, "step": 1960 }, { "epoch": 0.9231490159325211, - "grad_norm": 2.264636278152466, + "grad_norm": 2.4502975940704346, "learning_rate": 1.9423031865042177e-05, - "loss": 0.3329, + "loss": 0.4317, "step": 1970 }, { "epoch": 0.9278350515463918, - "grad_norm": 1.9120122194290161, + "grad_norm": 2.335707664489746, "learning_rate": 1.9420103092783507e-05, - "loss": 0.2805, + "loss": 0.405, "step": 1980 }, { "epoch": 0.9325210871602624, - "grad_norm": 2.0339694023132324, + "grad_norm": 2.0419764518737793, "learning_rate": 1.941717432052484e-05, - "loss": 0.3759, + "loss": 0.4669, "step": 1990 }, { "epoch": 0.9372071227741331, - "grad_norm": 1.7887804508209229, + "grad_norm": 2.2041420936584473, "learning_rate": 1.941424554826617e-05, - "loss": 0.3399, + "loss": 0.4696, "step": 2000 }, { "epoch": 0.9372071227741331, - "eval_loss": 0.05692083761096001, - "eval_pearson_cosine": 0.8112886737924754, - "eval_pearson_dot": 0.7242483960443309, - "eval_pearson_euclidean": 0.8065918617470516, - "eval_pearson_manhattan": 0.8072749119906675, - "eval_runtime": 4.6063, - "eval_samples_per_second": 325.643, - "eval_spearman_cosine": 0.8163229135143312, - "eval_spearman_dot": 0.722594730463142, - "eval_spearman_euclidean": 0.8151945311828125, - "eval_spearman_manhattan": 0.8162459677504973, - "eval_steps_per_second": 20.407, + "eval_loss": 0.10549145191907883, + "eval_pearson_cosine": 0.7617707450696116, + "eval_pearson_dot": 0.5909647410786931, + "eval_pearson_euclidean": 0.7526201710813609, + "eval_pearson_manhattan": 0.7514050722082715, + "eval_runtime": 3.4394, + "eval_samples_per_second": 436.124, + "eval_spearman_cosine": 0.7834635716420068, + "eval_spearman_dot": 0.5835376546734291, + "eval_spearman_euclidean": 0.7675113486625056, + "eval_spearman_manhattan": 0.7668717701506194, + "eval_steps_per_second": 27.33, "step": 2000 }, { "epoch": 0.9418931583880038, - "grad_norm": 1.454084873199463, + "grad_norm": 1.7690527439117432, "learning_rate": 1.94113167760075e-05, - "loss": 0.274, + "loss": 0.39, "step": 2010 }, { "epoch": 0.9465791940018744, - "grad_norm": 2.0930581092834473, + "grad_norm": 2.247870683670044, "learning_rate": 1.940838800374883e-05, - "loss": 0.3101, + "loss": 0.4283, "step": 2020 }, { "epoch": 0.9512652296157451, - "grad_norm": 2.347327947616577, + "grad_norm": 2.7865118980407715, "learning_rate": 1.940545923149016e-05, - "loss": 0.3134, + "loss": 0.4041, "step": 2030 }, { "epoch": 0.9559512652296157, - "grad_norm": 1.7699518203735352, + "grad_norm": 2.7173750400543213, "learning_rate": 1.9402530459231494e-05, - "loss": 0.2954, + "loss": 0.428, "step": 2040 }, { "epoch": 0.9606373008434864, - "grad_norm": 1.983152985572815, + "grad_norm": 2.4290664196014404, "learning_rate": 1.9399601686972823e-05, - "loss": 0.3267, + "loss": 0.4363, "step": 2050 }, { "epoch": 0.9653233364573571, - "grad_norm": 2.6133322715759277, + "grad_norm": 2.9478437900543213, "learning_rate": 1.9396672914714153e-05, - "loss": 0.3197, + "loss": 0.4341, "step": 2060 }, { "epoch": 0.9700093720712277, - "grad_norm": 1.6881980895996094, + "grad_norm": 2.5642406940460205, "learning_rate": 1.9393744142455482e-05, - "loss": 0.3284, + "loss": 0.4402, "step": 2070 }, { "epoch": 0.9746954076850984, - "grad_norm": 2.0169336795806885, + "grad_norm": 2.622969388961792, "learning_rate": 1.9390815370196815e-05, - "loss": 0.3245, + "loss": 0.4415, "step": 2080 }, { "epoch": 0.979381443298969, - "grad_norm": 1.456597924232483, + "grad_norm": 1.8864187002182007, "learning_rate": 1.9387886597938145e-05, - "loss": 0.297, + "loss": 0.421, "step": 2090 }, { "epoch": 0.9840674789128397, - "grad_norm": 1.7516179084777832, + "grad_norm": 2.145385503768921, "learning_rate": 1.9384957825679474e-05, - "loss": 0.3043, + "loss": 0.4142, "step": 2100 }, { "epoch": 0.9887535145267105, - "grad_norm": 1.981558084487915, + "grad_norm": 2.5870542526245117, "learning_rate": 1.9382029053420807e-05, - "loss": 0.3321, + "loss": 0.4701, "step": 2110 }, { "epoch": 0.993439550140581, - "grad_norm": 2.1244945526123047, + "grad_norm": 2.5176923274993896, "learning_rate": 1.9379100281162137e-05, - "loss": 0.2638, + "loss": 0.3776, "step": 2120 }, { "epoch": 0.9981255857544518, - "grad_norm": 2.025820016860962, + "grad_norm": 2.8260464668273926, "learning_rate": 1.937617150890347e-05, - "loss": 0.3164, + "loss": 0.4453, "step": 2130 }, { "epoch": 1.0028116213683225, - "grad_norm": 1.7136940956115723, + "grad_norm": 2.2735579013824463, "learning_rate": 1.93732427366448e-05, - "loss": 0.311, + "loss": 0.3617, "step": 2140 }, { "epoch": 1.007497656982193, - "grad_norm": 1.604434609413147, + "grad_norm": 1.9802652597427368, "learning_rate": 1.9370313964386132e-05, - "loss": 0.2511, + "loss": 0.3103, "step": 2150 }, { "epoch": 1.0121836925960637, - "grad_norm": 1.7466777563095093, + "grad_norm": 2.4149436950683594, "learning_rate": 1.936738519212746e-05, - "loss": 0.2342, + "loss": 0.3183, "step": 2160 }, { "epoch": 1.0168697282099344, - "grad_norm": 1.9601348638534546, + "grad_norm": 2.9166879653930664, "learning_rate": 1.9364456419868794e-05, - "loss": 0.2419, + "loss": 0.2927, "step": 2170 }, { "epoch": 1.021555763823805, - "grad_norm": 1.7632412910461426, + "grad_norm": 1.8774973154067993, "learning_rate": 1.9361527647610124e-05, - "loss": 0.2474, + "loss": 0.3217, "step": 2180 }, { "epoch": 1.0262417994376758, - "grad_norm": 1.8551344871520996, + "grad_norm": 2.0291717052459717, "learning_rate": 1.9358598875351457e-05, - "loss": 0.2375, + "loss": 0.3236, "step": 2190 }, { "epoch": 1.0309278350515463, - "grad_norm": 1.8569507598876953, + "grad_norm": 2.500614643096924, "learning_rate": 1.9355670103092786e-05, - "loss": 0.2226, + "loss": 0.3187, "step": 2200 }, { "epoch": 1.035613870665417, - "grad_norm": 1.5567635297775269, + "grad_norm": 2.4826183319091797, "learning_rate": 1.9352741330834116e-05, - "loss": 0.2615, + "loss": 0.3284, "step": 2210 }, { "epoch": 1.0402999062792877, - "grad_norm": 1.6635299921035767, + "grad_norm": 2.647724151611328, "learning_rate": 1.934981255857545e-05, - "loss": 0.255, + "loss": 0.3432, "step": 2220 }, { "epoch": 1.0449859418931584, - "grad_norm": 2.080941915512085, + "grad_norm": 2.356365203857422, "learning_rate": 1.9346883786316778e-05, - "loss": 0.2847, + "loss": 0.3321, "step": 2230 }, { "epoch": 1.0496719775070291, - "grad_norm": 1.8611458539962769, + "grad_norm": 2.514002561569214, "learning_rate": 1.9343955014058108e-05, - "loss": 0.2395, + "loss": 0.3115, "step": 2240 }, { "epoch": 1.0543580131208996, - "grad_norm": 1.7495652437210083, + "grad_norm": 2.049299716949463, "learning_rate": 1.9341026241799437e-05, - "loss": 0.2473, + "loss": 0.3474, "step": 2250 }, { "epoch": 1.0543580131208996, - "eval_loss": 0.045334625989198685, - "eval_pearson_cosine": 0.8123685468173392, - "eval_pearson_dot": 0.7270518178916561, - "eval_pearson_euclidean": 0.8020452343241686, - "eval_pearson_manhattan": 0.8030795563147421, - "eval_runtime": 5.4133, - "eval_samples_per_second": 277.096, - "eval_spearman_cosine": 0.814265717527882, - "eval_spearman_dot": 0.7261387412776578, - "eval_spearman_euclidean": 0.809310966263978, - "eval_spearman_manhattan": 0.8103368464209432, - "eval_steps_per_second": 17.365, + "eval_loss": 0.0818348377943039, + "eval_pearson_cosine": 0.7662947284133068, + "eval_pearson_dot": 0.5774155023095773, + "eval_pearson_euclidean": 0.7536499802604126, + "eval_pearson_manhattan": 0.7526750534069606, + "eval_runtime": 3.4518, + "eval_samples_per_second": 434.553, + "eval_spearman_cosine": 0.7776893657594519, + "eval_spearman_dot": 0.5747770364870878, + "eval_spearman_euclidean": 0.7641629894500831, + "eval_spearman_manhattan": 0.7635836296791149, + "eval_steps_per_second": 27.232, "step": 2250 }, { "epoch": 1.0590440487347703, - "grad_norm": 1.861383080482483, + "grad_norm": 1.9242613315582275, "learning_rate": 1.933809746954077e-05, - "loss": 0.2635, + "loss": 0.3362, "step": 2260 }, { "epoch": 1.063730084348641, - "grad_norm": 1.855754017829895, + "grad_norm": 2.6948235034942627, "learning_rate": 1.93351686972821e-05, - "loss": 0.2752, + "loss": 0.3524, "step": 2270 }, { "epoch": 1.0684161199625117, - "grad_norm": 1.7547088861465454, + "grad_norm": 2.162079334259033, "learning_rate": 1.9332239925023432e-05, - "loss": 0.2528, + "loss": 0.3292, "step": 2280 }, { "epoch": 1.0731021555763824, - "grad_norm": 2.317349672317505, + "grad_norm": 2.1351771354675293, "learning_rate": 1.9329311152764762e-05, - "loss": 0.2727, + "loss": 0.3409, "step": 2290 }, { "epoch": 1.077788191190253, - "grad_norm": 1.69141685962677, + "grad_norm": 2.527005434036255, "learning_rate": 1.932638238050609e-05, - "loss": 0.2576, + "loss": 0.3027, "step": 2300 }, { "epoch": 1.0824742268041236, - "grad_norm": 1.9386281967163086, + "grad_norm": 2.446911334991455, "learning_rate": 1.9323453608247424e-05, - "loss": 0.2746, + "loss": 0.3376, "step": 2310 }, { "epoch": 1.0871602624179943, - "grad_norm": 2.079533100128174, + "grad_norm": 2.478301763534546, "learning_rate": 1.9320524835988754e-05, - "loss": 0.2145, + "loss": 0.298, "step": 2320 }, { "epoch": 1.091846298031865, - "grad_norm": 2.5439393520355225, + "grad_norm": 2.537336587905884, "learning_rate": 1.9317596063730087e-05, - "loss": 0.2406, + "loss": 0.3312, "step": 2330 }, { "epoch": 1.0965323336457358, - "grad_norm": 2.132382869720459, + "grad_norm": 2.4405462741851807, "learning_rate": 1.9314667291471416e-05, - "loss": 0.2158, + "loss": 0.2998, "step": 2340 }, { "epoch": 1.1012183692596063, - "grad_norm": 2.181603193283081, + "grad_norm": 2.2255520820617676, "learning_rate": 1.931173851921275e-05, - "loss": 0.2358, + "loss": 0.2987, "step": 2350 }, { "epoch": 1.105904404873477, - "grad_norm": 2.288633108139038, + "grad_norm": 2.541879177093506, "learning_rate": 1.930880974695408e-05, - "loss": 0.2715, + "loss": 0.3286, "step": 2360 }, { "epoch": 1.1105904404873477, - "grad_norm": 1.882150650024414, + "grad_norm": 2.1010048389434814, "learning_rate": 1.930588097469541e-05, - "loss": 0.2111, + "loss": 0.2989, "step": 2370 }, { "epoch": 1.1152764761012184, - "grad_norm": 1.695289969444275, + "grad_norm": 1.9881564378738403, "learning_rate": 1.930295220243674e-05, - "loss": 0.2555, + "loss": 0.3537, "step": 2380 }, { "epoch": 1.119962511715089, - "grad_norm": 1.945979118347168, + "grad_norm": 2.503845691680908, "learning_rate": 1.930002343017807e-05, - "loss": 0.2415, + "loss": 0.3242, "step": 2390 }, { "epoch": 1.1246485473289598, - "grad_norm": 2.199965238571167, + "grad_norm": 2.773836612701416, "learning_rate": 1.9297094657919403e-05, - "loss": 0.2561, + "loss": 0.3144, "step": 2400 }, { "epoch": 1.1293345829428303, - "grad_norm": 2.0236504077911377, + "grad_norm": 2.7757749557495117, "learning_rate": 1.9294165885660733e-05, - "loss": 0.2458, + "loss": 0.3114, "step": 2410 }, { "epoch": 1.134020618556701, - "grad_norm": 1.7911378145217896, + "grad_norm": 2.707315444946289, "learning_rate": 1.9291237113402062e-05, - "loss": 0.2428, + "loss": 0.2872, "step": 2420 }, { "epoch": 1.1387066541705717, - "grad_norm": 1.8625153303146362, + "grad_norm": 2.1421258449554443, "learning_rate": 1.9288308341143395e-05, - "loss": 0.2482, + "loss": 0.3122, "step": 2430 }, { "epoch": 1.1433926897844424, - "grad_norm": 1.4607598781585693, + "grad_norm": 2.2792320251464844, "learning_rate": 1.9285379568884725e-05, - "loss": 0.2383, + "loss": 0.304, "step": 2440 }, { "epoch": 1.148078725398313, - "grad_norm": 1.552121877670288, + "grad_norm": 2.1971569061279297, "learning_rate": 1.9282450796626054e-05, - "loss": 0.2917, + "loss": 0.3571, "step": 2450 }, { "epoch": 1.1527647610121836, - "grad_norm": 2.049769163131714, + "grad_norm": 2.101923704147339, "learning_rate": 1.9279522024367387e-05, - "loss": 0.2241, + "loss": 0.3069, "step": 2460 }, { "epoch": 1.1574507966260543, - "grad_norm": 1.900490164756775, + "grad_norm": 2.395801305770874, "learning_rate": 1.9276593252108717e-05, - "loss": 0.2423, + "loss": 0.3145, "step": 2470 }, { "epoch": 1.162136832239925, - "grad_norm": 1.1432183980941772, + "grad_norm": 1.9261037111282349, "learning_rate": 1.927366447985005e-05, - "loss": 0.2556, + "loss": 0.3229, "step": 2480 }, { "epoch": 1.1668228678537957, - "grad_norm": 1.8964147567749023, + "grad_norm": 2.185063362121582, "learning_rate": 1.927073570759138e-05, - "loss": 0.2258, + "loss": 0.3289, "step": 2490 }, { "epoch": 1.1715089034676662, - "grad_norm": 1.8627080917358398, + "grad_norm": 2.2372336387634277, "learning_rate": 1.926780693533271e-05, - "loss": 0.2563, + "loss": 0.319, "step": 2500 }, { "epoch": 1.1715089034676662, - "eval_loss": 0.04076731204986572, - "eval_pearson_cosine": 0.8177640807226254, - "eval_pearson_dot": 0.7517947588072431, - "eval_pearson_euclidean": 0.8032143754633552, - "eval_pearson_manhattan": 0.804324819579108, - "eval_runtime": 4.5167, - "eval_samples_per_second": 332.103, - "eval_spearman_cosine": 0.8194654831329402, - "eval_spearman_dot": 0.7504183673719423, - "eval_spearman_euclidean": 0.811997591732405, - "eval_spearman_manhattan": 0.8131700838529736, - "eval_steps_per_second": 20.812, + "eval_loss": 0.07521134614944458, + "eval_pearson_cosine": 0.7753374743054451, + "eval_pearson_dot": 0.5929128805152892, + "eval_pearson_euclidean": 0.7592332760910097, + "eval_pearson_manhattan": 0.7589109467518398, + "eval_runtime": 3.6686, + "eval_samples_per_second": 408.877, + "eval_spearman_cosine": 0.785757446712241, + "eval_spearman_dot": 0.5918750155733082, + "eval_spearman_euclidean": 0.7692376569048597, + "eval_spearman_manhattan": 0.7692457770843739, + "eval_steps_per_second": 25.623, "step": 2500 }, { "epoch": 1.176194939081537, - "grad_norm": 1.7957406044006348, + "grad_norm": 2.2003061771392822, "learning_rate": 1.926487816307404e-05, - "loss": 0.2381, + "loss": 0.2979, "step": 2510 }, { "epoch": 1.1808809746954076, - "grad_norm": 1.868900179862976, + "grad_norm": 1.9160058498382568, "learning_rate": 1.926194939081537e-05, - "loss": 0.245, + "loss": 0.3119, "step": 2520 }, { "epoch": 1.1855670103092784, - "grad_norm": 1.8972773551940918, + "grad_norm": 2.5396270751953125, "learning_rate": 1.9259020618556704e-05, - "loss": 0.2346, + "loss": 0.2882, "step": 2530 }, { "epoch": 1.190253045923149, - "grad_norm": 1.7982817888259888, + "grad_norm": 2.184056520462036, "learning_rate": 1.9256091846298033e-05, - "loss": 0.2182, + "loss": 0.3012, "step": 2540 }, { "epoch": 1.1949390815370198, - "grad_norm": 2.066537380218506, + "grad_norm": 2.4930028915405273, "learning_rate": 1.9253163074039366e-05, - "loss": 0.2396, + "loss": 0.3399, "step": 2550 }, { "epoch": 1.1996251171508903, - "grad_norm": 2.3115556240081787, + "grad_norm": 2.0577898025512695, "learning_rate": 1.9250234301780696e-05, - "loss": 0.2513, + "loss": 0.3013, "step": 2560 }, { "epoch": 1.204311152764761, - "grad_norm": 2.140615940093994, + "grad_norm": 1.9735450744628906, "learning_rate": 1.9247305529522025e-05, - "loss": 0.258, + "loss": 0.315, "step": 2570 }, { "epoch": 1.2089971883786317, - "grad_norm": 1.9183828830718994, + "grad_norm": 2.815910577774048, "learning_rate": 1.9244376757263358e-05, - "loss": 0.2249, + "loss": 0.3134, "step": 2580 }, { "epoch": 1.2136832239925024, - "grad_norm": 1.9000664949417114, + "grad_norm": 2.4166650772094727, "learning_rate": 1.9241447985004687e-05, - "loss": 0.2469, + "loss": 0.2924, "step": 2590 }, { "epoch": 1.218369259606373, - "grad_norm": 2.2240567207336426, + "grad_norm": 2.65451979637146, "learning_rate": 1.9238519212746017e-05, - "loss": 0.2342, + "loss": 0.313, "step": 2600 }, { "epoch": 1.2230552952202436, - "grad_norm": 2.227921724319458, + "grad_norm": 3.048393726348877, "learning_rate": 1.923559044048735e-05, - "loss": 0.2867, + "loss": 0.3396, "step": 2610 }, { "epoch": 1.2277413308341143, - "grad_norm": 1.8266960382461548, + "grad_norm": 2.1255080699920654, "learning_rate": 1.923266166822868e-05, - "loss": 0.2541, + "loss": 0.2918, "step": 2620 }, { "epoch": 1.232427366447985, - "grad_norm": 2.0011913776397705, + "grad_norm": 2.4257168769836426, "learning_rate": 1.922973289597001e-05, - "loss": 0.2404, + "loss": 0.3221, "step": 2630 }, { "epoch": 1.2371134020618557, - "grad_norm": 2.053349256515503, + "grad_norm": 2.3977224826812744, "learning_rate": 1.9226804123711342e-05, - "loss": 0.2931, + "loss": 0.3777, "step": 2640 }, { "epoch": 1.2417994376757264, - "grad_norm": 1.3468660116195679, + "grad_norm": 2.0743911266326904, "learning_rate": 1.922387535145267e-05, - "loss": 0.2326, + "loss": 0.2981, "step": 2650 }, { "epoch": 1.246485473289597, - "grad_norm": 2.1676151752471924, + "grad_norm": 2.740622043609619, "learning_rate": 1.9220946579194004e-05, - "loss": 0.2478, + "loss": 0.3239, "step": 2660 }, { "epoch": 1.2511715089034676, - "grad_norm": 1.9708117246627808, + "grad_norm": 2.5174942016601562, "learning_rate": 1.9218017806935334e-05, - "loss": 0.2281, + "loss": 0.2981, "step": 2670 }, { "epoch": 1.2558575445173383, - "grad_norm": 1.8341456651687622, + "grad_norm": 2.3390488624572754, "learning_rate": 1.9215089034676667e-05, - "loss": 0.2477, + "loss": 0.3509, "step": 2680 }, { "epoch": 1.260543580131209, - "grad_norm": 1.9229050874710083, + "grad_norm": 2.495445966720581, "learning_rate": 1.9212160262417996e-05, - "loss": 0.2424, + "loss": 0.3153, "step": 2690 }, { "epoch": 1.2652296157450795, - "grad_norm": 2.013848066329956, + "grad_norm": 2.384673595428467, "learning_rate": 1.9209231490159326e-05, - "loss": 0.272, + "loss": 0.3541, "step": 2700 }, { "epoch": 1.2699156513589505, - "grad_norm": 1.684461236000061, + "grad_norm": 1.9707324504852295, "learning_rate": 1.920630271790066e-05, - "loss": 0.2432, + "loss": 0.3053, "step": 2710 }, { "epoch": 1.274601686972821, - "grad_norm": 1.300451636314392, + "grad_norm": 2.0584781169891357, "learning_rate": 1.9203373945641988e-05, - "loss": 0.2006, + "loss": 0.2716, "step": 2720 }, { "epoch": 1.2792877225866917, - "grad_norm": 1.6482150554656982, + "grad_norm": 1.9361029863357544, "learning_rate": 1.920044517338332e-05, - "loss": 0.2495, + "loss": 0.2988, "step": 2730 }, { "epoch": 1.2839737582005624, - "grad_norm": 1.7251955270767212, + "grad_norm": 2.3929100036621094, "learning_rate": 1.919751640112465e-05, - "loss": 0.2469, + "loss": 0.3356, "step": 2740 }, { "epoch": 1.2886597938144329, - "grad_norm": 2.6221530437469482, + "grad_norm": 2.7867496013641357, "learning_rate": 1.9194587628865983e-05, - "loss": 0.2841, + "loss": 0.3682, "step": 2750 }, { "epoch": 1.2886597938144329, - "eval_loss": 0.04372455179691315, - "eval_pearson_cosine": 0.8073759135651102, - "eval_pearson_dot": 0.7236907127152534, - "eval_pearson_euclidean": 0.8053106447795528, - "eval_pearson_manhattan": 0.806321043432086, - "eval_runtime": 3.4277, - "eval_samples_per_second": 437.612, - "eval_spearman_cosine": 0.8100379018669143, - "eval_spearman_dot": 0.7204492861182332, - "eval_spearman_euclidean": 0.8129520365277123, - "eval_spearman_manhattan": 0.813834648092825, - "eval_steps_per_second": 27.424, + "eval_loss": 0.07667941600084305, + "eval_pearson_cosine": 0.7735852714798206, + "eval_pearson_dot": 0.578414803552846, + "eval_pearson_euclidean": 0.7564370361137946, + "eval_pearson_manhattan": 0.7556229015098381, + "eval_runtime": 3.3395, + "eval_samples_per_second": 449.165, + "eval_spearman_cosine": 0.7851408708188664, + "eval_spearman_dot": 0.5785312770170487, + "eval_spearman_euclidean": 0.7670951414033645, + "eval_spearman_manhattan": 0.7667469763884154, + "eval_steps_per_second": 28.148, "step": 2750 }, { "epoch": 1.2933458294283038, - "grad_norm": 1.8232098817825317, + "grad_norm": 2.0457334518432617, "learning_rate": 1.9191658856607313e-05, - "loss": 0.2717, + "loss": 0.3118, "step": 2760 }, { "epoch": 1.2980318650421743, - "grad_norm": 1.6222856044769287, + "grad_norm": 2.1723532676696777, "learning_rate": 1.9188730084348642e-05, - "loss": 0.227, + "loss": 0.2918, "step": 2770 }, { "epoch": 1.302717900656045, - "grad_norm": 1.8859031200408936, + "grad_norm": 2.041738748550415, "learning_rate": 1.9185801312089972e-05, - "loss": 0.2428, + "loss": 0.3084, "step": 2780 }, { "epoch": 1.3074039362699157, - "grad_norm": 1.9047764539718628, + "grad_norm": 2.774698257446289, "learning_rate": 1.9182872539831305e-05, - "loss": 0.2322, + "loss": 0.3337, "step": 2790 }, { "epoch": 1.3120899718837864, - "grad_norm": 1.734043002128601, + "grad_norm": 1.9762837886810303, "learning_rate": 1.9179943767572634e-05, - "loss": 0.2562, + "loss": 0.3029, "step": 2800 }, { "epoch": 1.316776007497657, - "grad_norm": 1.840468406677246, + "grad_norm": 2.1785385608673096, "learning_rate": 1.9177014995313964e-05, - "loss": 0.2429, + "loss": 0.3337, "step": 2810 }, { "epoch": 1.3214620431115276, - "grad_norm": 1.6314520835876465, + "grad_norm": 2.0208494663238525, "learning_rate": 1.9174086223055296e-05, - "loss": 0.232, + "loss": 0.2931, "step": 2820 }, { "epoch": 1.3261480787253983, - "grad_norm": 1.9515080451965332, + "grad_norm": 2.3976972103118896, "learning_rate": 1.9171157450796626e-05, - "loss": 0.2289, + "loss": 0.3316, "step": 2830 }, { "epoch": 1.330834114339269, - "grad_norm": 1.7130558490753174, + "grad_norm": 1.9782285690307617, "learning_rate": 1.916822867853796e-05, - "loss": 0.2349, + "loss": 0.3173, "step": 2840 }, { "epoch": 1.3355201499531397, - "grad_norm": 2.0093345642089844, + "grad_norm": 2.0938830375671387, "learning_rate": 1.916529990627929e-05, - "loss": 0.216, + "loss": 0.2858, "step": 2850 }, { "epoch": 1.3402061855670104, - "grad_norm": 1.8628073930740356, + "grad_norm": 2.079247236251831, "learning_rate": 1.916237113402062e-05, - "loss": 0.2395, + "loss": 0.2911, "step": 2860 }, { "epoch": 1.344892221180881, - "grad_norm": 1.8130676746368408, + "grad_norm": 2.5487163066864014, "learning_rate": 1.915944236176195e-05, - "loss": 0.2283, + "loss": 0.2782, "step": 2870 }, { "epoch": 1.3495782567947516, - "grad_norm": 1.8690656423568726, + "grad_norm": 2.4930672645568848, "learning_rate": 1.9156513589503284e-05, - "loss": 0.2459, + "loss": 0.309, "step": 2880 }, { "epoch": 1.3542642924086223, - "grad_norm": 1.607759714126587, + "grad_norm": 2.208987236022949, "learning_rate": 1.9153584817244613e-05, - "loss": 0.2128, + "loss": 0.302, "step": 2890 }, { "epoch": 1.358950328022493, - "grad_norm": 1.827553629875183, + "grad_norm": 2.2777154445648193, "learning_rate": 1.9150656044985943e-05, - "loss": 0.225, + "loss": 0.2932, "step": 2900 }, { "epoch": 1.3636363636363638, - "grad_norm": 1.599199891090393, + "grad_norm": 2.0704426765441895, "learning_rate": 1.9147727272727276e-05, - "loss": 0.264, + "loss": 0.3326, "step": 2910 }, { "epoch": 1.3683223992502342, - "grad_norm": 2.0378599166870117, + "grad_norm": 2.550323486328125, "learning_rate": 1.9144798500468605e-05, - "loss": 0.2147, + "loss": 0.2945, "step": 2920 }, { "epoch": 1.373008434864105, - "grad_norm": 2.201847791671753, + "grad_norm": 2.1471896171569824, "learning_rate": 1.9141869728209938e-05, - "loss": 0.2812, + "loss": 0.3258, "step": 2930 }, { "epoch": 1.3776944704779757, - "grad_norm": 1.5552937984466553, + "grad_norm": 2.4434566497802734, "learning_rate": 1.9138940955951267e-05, - "loss": 0.2386, + "loss": 0.3, "step": 2940 }, { "epoch": 1.3823805060918464, - "grad_norm": 1.7810213565826416, + "grad_norm": 2.2125484943389893, "learning_rate": 1.9136012183692597e-05, - "loss": 0.2596, + "loss": 0.3618, "step": 2950 }, { "epoch": 1.387066541705717, - "grad_norm": 1.9728736877441406, + "grad_norm": 2.5156352519989014, "learning_rate": 1.913308341143393e-05, - "loss": 0.2385, + "loss": 0.3102, "step": 2960 }, { "epoch": 1.3917525773195876, - "grad_norm": 1.868575930595398, + "grad_norm": 2.1281540393829346, "learning_rate": 1.913015463917526e-05, - "loss": 0.2287, + "loss": 0.31, "step": 2970 }, { "epoch": 1.3964386129334583, - "grad_norm": 1.4674561023712158, + "grad_norm": 2.1260693073272705, "learning_rate": 1.912722586691659e-05, - "loss": 0.2619, + "loss": 0.2946, "step": 2980 }, { "epoch": 1.401124648547329, - "grad_norm": 1.9503545761108398, + "grad_norm": 2.591770887374878, "learning_rate": 1.912429709465792e-05, - "loss": 0.2319, + "loss": 0.2906, "step": 2990 }, { "epoch": 1.4058106841611997, - "grad_norm": 1.705707311630249, + "grad_norm": 1.832505226135254, "learning_rate": 1.912136832239925e-05, - "loss": 0.2462, + "loss": 0.3033, "step": 3000 }, { "epoch": 1.4058106841611997, - "eval_loss": 0.04194045811891556, - "eval_pearson_cosine": 0.8164473390079152, - "eval_pearson_dot": 0.7394876158890895, - "eval_pearson_euclidean": 0.8038670284436478, - "eval_pearson_manhattan": 0.8050459887967918, - "eval_runtime": 3.394, - "eval_samples_per_second": 441.953, - "eval_spearman_cosine": 0.8192114591290918, - "eval_spearman_dot": 0.7393194455764386, - "eval_spearman_euclidean": 0.8132475290397445, - "eval_spearman_manhattan": 0.8142678938453525, - "eval_steps_per_second": 27.696, + "eval_loss": 0.07158821821212769, + "eval_pearson_cosine": 0.7835729367892306, + "eval_pearson_dot": 0.5987220055863389, + "eval_pearson_euclidean": 0.7599913918565986, + "eval_pearson_manhattan": 0.7589542766532986, + "eval_runtime": 3.4365, + "eval_samples_per_second": 436.486, + "eval_spearman_cosine": 0.7961581681624044, + "eval_spearman_dot": 0.5975812797826083, + "eval_spearman_euclidean": 0.7727385023345612, + "eval_spearman_manhattan": 0.7723228512152759, + "eval_steps_per_second": 27.353, "step": 3000 }, { "epoch": 1.4104967197750704, - "grad_norm": 2.0516934394836426, + "grad_norm": 2.3309245109558105, "learning_rate": 1.911843955014058e-05, - "loss": 0.2469, + "loss": 0.3238, "step": 3010 }, { "epoch": 1.415182755388941, - "grad_norm": 1.8646856546401978, + "grad_norm": 2.3359642028808594, "learning_rate": 1.9115510777881914e-05, - "loss": 0.2434, + "loss": 0.2933, "step": 3020 }, { "epoch": 1.4198687910028116, - "grad_norm": 1.746596097946167, + "grad_norm": 2.363147497177124, "learning_rate": 1.9112582005623243e-05, - "loss": 0.2392, + "loss": 0.3204, "step": 3030 }, { "epoch": 1.4245548266166823, - "grad_norm": 1.7546746730804443, + "grad_norm": 2.353273391723633, "learning_rate": 1.9109653233364576e-05, - "loss": 0.2141, + "loss": 0.295, "step": 3040 }, { "epoch": 1.429240862230553, - "grad_norm": 1.7862520217895508, + "grad_norm": 2.418267011642456, "learning_rate": 1.9106724461105905e-05, - "loss": 0.2424, + "loss": 0.3397, "step": 3050 }, { "epoch": 1.4339268978444237, - "grad_norm": 2.089353322982788, + "grad_norm": 2.612607955932617, "learning_rate": 1.910379568884724e-05, - "loss": 0.2493, + "loss": 0.3001, "step": 3060 }, { "epoch": 1.4386129334582942, - "grad_norm": 1.7390618324279785, + "grad_norm": 2.506488084793091, "learning_rate": 1.9100866916588568e-05, - "loss": 0.2317, + "loss": 0.3101, "step": 3070 }, { "epoch": 1.443298969072165, - "grad_norm": 1.9297877550125122, + "grad_norm": 2.437333345413208, "learning_rate": 1.90979381443299e-05, - "loss": 0.2335, + "loss": 0.3017, "step": 3080 }, { "epoch": 1.4479850046860356, - "grad_norm": 1.872578740119934, + "grad_norm": 2.3391661643981934, "learning_rate": 1.909500937207123e-05, - "loss": 0.222, + "loss": 0.2943, "step": 3090 }, { "epoch": 1.4526710402999063, - "grad_norm": 1.646843433380127, + "grad_norm": 2.410454511642456, "learning_rate": 1.909208059981256e-05, - "loss": 0.2154, + "loss": 0.3242, "step": 3100 }, { "epoch": 1.457357075913777, - "grad_norm": 1.8208719491958618, + "grad_norm": 2.6086418628692627, "learning_rate": 1.9089151827553893e-05, - "loss": 0.2887, + "loss": 0.3442, "step": 3110 }, { "epoch": 1.4620431115276475, - "grad_norm": 1.6147174835205078, + "grad_norm": 2.1769497394561768, "learning_rate": 1.9086223055295222e-05, - "loss": 0.2292, + "loss": 0.3114, "step": 3120 }, { "epoch": 1.4667291471415183, - "grad_norm": 2.147585391998291, + "grad_norm": 1.9582467079162598, "learning_rate": 1.908329428303655e-05, - "loss": 0.2598, + "loss": 0.3261, "step": 3130 }, { "epoch": 1.471415182755389, - "grad_norm": 2.17818284034729, + "grad_norm": 2.577211380004883, "learning_rate": 1.9080365510777884e-05, - "loss": 0.2177, + "loss": 0.31, "step": 3140 }, { "epoch": 1.4761012183692597, - "grad_norm": 1.7800393104553223, + "grad_norm": 2.725773811340332, "learning_rate": 1.9077436738519214e-05, - "loss": 0.2091, + "loss": 0.2859, "step": 3150 }, { "epoch": 1.4807872539831304, - "grad_norm": 1.8269144296646118, + "grad_norm": 2.779578685760498, "learning_rate": 1.9074507966260543e-05, - "loss": 0.241, + "loss": 0.3304, "step": 3160 }, { "epoch": 1.4854732895970009, - "grad_norm": 1.4544728994369507, + "grad_norm": 1.9493879079818726, "learning_rate": 1.9071579194001876e-05, - "loss": 0.2481, + "loss": 0.2992, "step": 3170 }, { "epoch": 1.4901593252108716, - "grad_norm": 2.5941991806030273, + "grad_norm": 3.1011288166046143, "learning_rate": 1.9068650421743206e-05, - "loss": 0.266, + "loss": 0.3406, "step": 3180 }, { "epoch": 1.4948453608247423, - "grad_norm": 1.9068180322647095, + "grad_norm": 2.466146230697632, "learning_rate": 1.906572164948454e-05, - "loss": 0.2546, + "loss": 0.357, "step": 3190 }, { "epoch": 1.499531396438613, - "grad_norm": 1.590065598487854, + "grad_norm": 2.0382285118103027, "learning_rate": 1.9062792877225868e-05, - "loss": 0.2292, + "loss": 0.2968, "step": 3200 }, { "epoch": 1.5042174320524837, - "grad_norm": 1.9283113479614258, + "grad_norm": 2.331540584564209, "learning_rate": 1.9059864104967198e-05, - "loss": 0.2485, + "loss": 0.3125, "step": 3210 }, { "epoch": 1.5089034676663542, - "grad_norm": 2.1952388286590576, + "grad_norm": 2.590193271636963, "learning_rate": 1.905693533270853e-05, - "loss": 0.2197, + "loss": 0.2762, "step": 3220 }, { "epoch": 1.513589503280225, - "grad_norm": 1.554611086845398, + "grad_norm": 2.878938913345337, "learning_rate": 1.905400656044986e-05, - "loss": 0.2136, + "loss": 0.2868, "step": 3230 }, { "epoch": 1.5182755388940956, - "grad_norm": 1.5786999464035034, + "grad_norm": 2.4432168006896973, "learning_rate": 1.9051077788191193e-05, - "loss": 0.2333, + "loss": 0.3021, "step": 3240 }, { "epoch": 1.522961574507966, - "grad_norm": 2.7858917713165283, + "grad_norm": 2.849559783935547, "learning_rate": 1.9048149015932523e-05, - "loss": 0.2328, + "loss": 0.3247, "step": 3250 }, { "epoch": 1.522961574507966, - "eval_loss": 0.04040240868926048, - "eval_pearson_cosine": 0.8187130517711054, - "eval_pearson_dot": 0.7426107734380025, - "eval_pearson_euclidean": 0.8069722435152613, - "eval_pearson_manhattan": 0.8083599378238802, - "eval_runtime": 3.6772, - "eval_samples_per_second": 407.917, - "eval_spearman_cosine": 0.8202782064729509, - "eval_spearman_dot": 0.7414284976652127, - "eval_spearman_euclidean": 0.8153801634076588, - "eval_spearman_manhattan": 0.8165145388144227, - "eval_steps_per_second": 25.563, + "eval_loss": 0.07676094025373459, + "eval_pearson_cosine": 0.7779113850636818, + "eval_pearson_dot": 0.5637961042064976, + "eval_pearson_euclidean": 0.7621398908224677, + "eval_pearson_manhattan": 0.7612994290780648, + "eval_runtime": 3.2908, + "eval_samples_per_second": 455.817, + "eval_spearman_cosine": 0.7910676412952927, + "eval_spearman_dot": 0.562324725935799, + "eval_spearman_euclidean": 0.7734903614875008, + "eval_spearman_manhattan": 0.7731474604573987, + "eval_steps_per_second": 28.565, "step": 3250 }, { "epoch": 1.527647610121837, - "grad_norm": 1.9804577827453613, + "grad_norm": 2.372697591781616, "learning_rate": 1.9045220243673855e-05, - "loss": 0.252, + "loss": 0.3423, "step": 3260 }, { "epoch": 1.5323336457357075, - "grad_norm": 1.8716363906860352, + "grad_norm": 2.9068384170532227, "learning_rate": 1.9042291471415185e-05, - "loss": 0.2266, + "loss": 0.3133, "step": 3270 }, { "epoch": 1.5370196813495782, - "grad_norm": 1.834250569343567, + "grad_norm": 2.738391160964966, "learning_rate": 1.9039362699156518e-05, - "loss": 0.2314, + "loss": 0.3018, "step": 3280 }, { "epoch": 1.541705716963449, - "grad_norm": 2.0380077362060547, + "grad_norm": 2.121777296066284, "learning_rate": 1.9036433926897847e-05, - "loss": 0.2545, + "loss": 0.3203, "step": 3290 }, { "epoch": 1.5463917525773194, - "grad_norm": 1.8950886726379395, + "grad_norm": 2.1887810230255127, "learning_rate": 1.9033505154639177e-05, - "loss": 0.2902, + "loss": 0.3639, "step": 3300 }, { "epoch": 1.5510777881911904, - "grad_norm": 1.4216679334640503, + "grad_norm": 2.4203438758850098, "learning_rate": 1.9030576382380506e-05, - "loss": 0.2344, + "loss": 0.3116, "step": 3310 }, { "epoch": 1.5557638238050608, - "grad_norm": 1.2056218385696411, + "grad_norm": 2.00972318649292, "learning_rate": 1.902764761012184e-05, - "loss": 0.252, + "loss": 0.3139, "step": 3320 }, { "epoch": 1.5604498594189316, - "grad_norm": 1.8112496137619019, + "grad_norm": 2.093622922897339, "learning_rate": 1.902471883786317e-05, - "loss": 0.2406, + "loss": 0.3248, "step": 3330 }, { "epoch": 1.5651358950328023, - "grad_norm": 1.9375700950622559, + "grad_norm": 2.5851595401763916, "learning_rate": 1.9021790065604498e-05, - "loss": 0.2728, + "loss": 0.3386, "step": 3340 }, { "epoch": 1.569821930646673, - "grad_norm": 2.4203391075134277, + "grad_norm": 2.208651304244995, "learning_rate": 1.901886129334583e-05, - "loss": 0.2434, + "loss": 0.3138, "step": 3350 }, { "epoch": 1.5745079662605437, - "grad_norm": 2.114474058151245, + "grad_norm": 2.5663788318634033, "learning_rate": 1.901593252108716e-05, - "loss": 0.2535, + "loss": 0.3165, "step": 3360 }, { "epoch": 1.5791940018744142, - "grad_norm": 1.5152201652526855, + "grad_norm": 2.443042039871216, "learning_rate": 1.9013003748828493e-05, - "loss": 0.2537, + "loss": 0.3178, "step": 3370 }, { "epoch": 1.5838800374882849, - "grad_norm": 1.8051055669784546, + "grad_norm": 2.385019540786743, "learning_rate": 1.9010074976569823e-05, - "loss": 0.2271, + "loss": 0.2779, "step": 3380 }, { "epoch": 1.5885660731021556, - "grad_norm": 2.005028247833252, + "grad_norm": 2.5440433025360107, "learning_rate": 1.9007146204311156e-05, - "loss": 0.2508, + "loss": 0.3182, "step": 3390 }, { "epoch": 1.5932521087160263, - "grad_norm": 1.534379005432129, + "grad_norm": 2.1016016006469727, "learning_rate": 1.9004217432052485e-05, - "loss": 0.2358, + "loss": 0.3224, "step": 3400 }, { "epoch": 1.597938144329897, - "grad_norm": 1.2152713537216187, + "grad_norm": 1.8776601552963257, "learning_rate": 1.9001288659793815e-05, - "loss": 0.203, + "loss": 0.2926, "step": 3410 }, { "epoch": 1.6026241799437675, - "grad_norm": 1.584352970123291, + "grad_norm": 1.7855247259140015, "learning_rate": 1.8998359887535148e-05, - "loss": 0.2369, + "loss": 0.3087, "step": 3420 }, { "epoch": 1.6073102155576382, - "grad_norm": 1.8603837490081787, + "grad_norm": 2.239816188812256, "learning_rate": 1.8995431115276477e-05, - "loss": 0.2576, + "loss": 0.3254, "step": 3430 }, { "epoch": 1.611996251171509, - "grad_norm": 1.300493597984314, + "grad_norm": 2.309274673461914, "learning_rate": 1.899250234301781e-05, - "loss": 0.2048, + "loss": 0.2741, "step": 3440 }, { "epoch": 1.6166822867853796, - "grad_norm": 1.6629600524902344, + "grad_norm": 2.672464370727539, "learning_rate": 1.898957357075914e-05, - "loss": 0.2305, + "loss": 0.3137, "step": 3450 }, { "epoch": 1.6213683223992503, - "grad_norm": 2.555297374725342, + "grad_norm": 2.3761508464813232, "learning_rate": 1.8986644798500473e-05, - "loss": 0.2487, + "loss": 0.322, "step": 3460 }, { "epoch": 1.6260543580131208, - "grad_norm": 2.425975799560547, + "grad_norm": 2.497772216796875, "learning_rate": 1.8983716026241802e-05, - "loss": 0.241, + "loss": 0.2832, "step": 3470 }, { "epoch": 1.6307403936269915, - "grad_norm": 1.912858247756958, + "grad_norm": 2.2633016109466553, "learning_rate": 1.898078725398313e-05, - "loss": 0.2074, + "loss": 0.2659, "step": 3480 }, { "epoch": 1.6354264292408622, - "grad_norm": 2.066469430923462, + "grad_norm": 2.2554588317871094, "learning_rate": 1.897785848172446e-05, - "loss": 0.2157, + "loss": 0.3039, "step": 3490 }, { "epoch": 1.640112464854733, - "grad_norm": 1.5405519008636475, + "grad_norm": 2.0419113636016846, "learning_rate": 1.8974929709465794e-05, - "loss": 0.2052, + "loss": 0.26, "step": 3500 }, { "epoch": 1.640112464854733, - "eval_loss": 0.03896905109286308, - "eval_pearson_cosine": 0.8147239668269464, - "eval_pearson_dot": 0.74259279719071, - "eval_pearson_euclidean": 0.803459120860885, - "eval_pearson_manhattan": 0.8045289290680273, - "eval_runtime": 3.5517, - "eval_samples_per_second": 422.338, - "eval_spearman_cosine": 0.8164243711017356, - "eval_spearman_dot": 0.742209739118843, - "eval_spearman_euclidean": 0.8121850450179654, - "eval_spearman_manhattan": 0.8128758362528613, - "eval_steps_per_second": 26.467, + "eval_loss": 0.06864778697490692, + "eval_pearson_cosine": 0.7792468184875077, + "eval_pearson_dot": 0.6003709070725307, + "eval_pearson_euclidean": 0.762291216964627, + "eval_pearson_manhattan": 0.7615198846806042, + "eval_runtime": 3.2918, + "eval_samples_per_second": 455.681, + "eval_spearman_cosine": 0.7902077133733839, + "eval_spearman_dot": 0.5998471986095437, + "eval_spearman_euclidean": 0.7733862684662854, + "eval_spearman_manhattan": 0.7732713082323596, + "eval_steps_per_second": 28.556, "step": 3500 }, { "epoch": 1.6447985004686037, - "grad_norm": 1.4605026245117188, + "grad_norm": 2.1808488368988037, "learning_rate": 1.8972000937207123e-05, - "loss": 0.2329, + "loss": 0.2977, "step": 3510 }, { "epoch": 1.6494845360824741, - "grad_norm": 1.496071457862854, + "grad_norm": 2.5023481845855713, "learning_rate": 1.8969072164948453e-05, - "loss": 0.2171, + "loss": 0.3275, "step": 3520 }, { "epoch": 1.6541705716963448, - "grad_norm": 1.5330686569213867, + "grad_norm": 1.9074852466583252, "learning_rate": 1.8966143392689786e-05, - "loss": 0.235, + "loss": 0.3084, "step": 3530 }, { "epoch": 1.6588566073102156, - "grad_norm": 1.77309250831604, + "grad_norm": 2.4666028022766113, "learning_rate": 1.8963214620431115e-05, - "loss": 0.2398, + "loss": 0.3202, "step": 3540 }, { "epoch": 1.6635426429240863, - "grad_norm": 2.6333131790161133, + "grad_norm": 2.2712929248809814, "learning_rate": 1.8960285848172448e-05, - "loss": 0.2444, + "loss": 0.3023, "step": 3550 }, { "epoch": 1.668228678537957, - "grad_norm": 1.7444895505905151, + "grad_norm": 2.274965286254883, "learning_rate": 1.8957357075913778e-05, - "loss": 0.2308, + "loss": 0.3134, "step": 3560 }, { "epoch": 1.6729147141518275, - "grad_norm": 1.9555528163909912, + "grad_norm": 2.2933590412139893, "learning_rate": 1.895442830365511e-05, - "loss": 0.2177, + "loss": 0.2935, "step": 3570 }, { "epoch": 1.6776007497656982, - "grad_norm": 1.8581876754760742, + "grad_norm": 2.492459774017334, "learning_rate": 1.895149953139644e-05, - "loss": 0.2287, + "loss": 0.2891, "step": 3580 }, { "epoch": 1.6822867853795689, - "grad_norm": 1.9251590967178345, + "grad_norm": 2.4401142597198486, "learning_rate": 1.894857075913777e-05, - "loss": 0.2547, + "loss": 0.329, "step": 3590 }, { "epoch": 1.6869728209934396, - "grad_norm": 1.6071839332580566, + "grad_norm": 2.0816047191619873, "learning_rate": 1.8945641986879102e-05, - "loss": 0.2409, + "loss": 0.3042, "step": 3600 }, { "epoch": 1.6916588566073103, - "grad_norm": 1.735236406326294, + "grad_norm": 2.2334861755371094, "learning_rate": 1.8942713214620432e-05, - "loss": 0.2315, + "loss": 0.2909, "step": 3610 }, { "epoch": 1.6963448922211808, - "grad_norm": 1.5945345163345337, + "grad_norm": 2.3555166721343994, "learning_rate": 1.8939784442361765e-05, - "loss": 0.2283, + "loss": 0.3284, "step": 3620 }, { "epoch": 1.7010309278350515, - "grad_norm": 1.8697439432144165, + "grad_norm": 1.751710295677185, "learning_rate": 1.8936855670103094e-05, - "loss": 0.2382, + "loss": 0.2848, "step": 3630 }, { "epoch": 1.7057169634489222, - "grad_norm": 1.9775031805038452, + "grad_norm": 1.920424461364746, "learning_rate": 1.8933926897844427e-05, - "loss": 0.2167, + "loss": 0.279, "step": 3640 }, { "epoch": 1.710402999062793, - "grad_norm": 1.7976644039154053, + "grad_norm": 2.840911865234375, "learning_rate": 1.8930998125585757e-05, - "loss": 0.2252, + "loss": 0.3116, "step": 3650 }, { "epoch": 1.7150890346766636, - "grad_norm": 1.9334297180175781, + "grad_norm": 2.429884672164917, "learning_rate": 1.8928069353327086e-05, - "loss": 0.207, + "loss": 0.264, "step": 3660 }, { "epoch": 1.7197750702905341, - "grad_norm": 1.661363124847412, + "grad_norm": 2.2484772205352783, "learning_rate": 1.892514058106842e-05, - "loss": 0.2531, + "loss": 0.3201, "step": 3670 }, { "epoch": 1.7244611059044048, - "grad_norm": 2.2192280292510986, + "grad_norm": 2.342460870742798, "learning_rate": 1.892221180880975e-05, - "loss": 0.2552, + "loss": 0.3068, "step": 3680 }, { "epoch": 1.7291471415182755, - "grad_norm": 1.9910961389541626, + "grad_norm": 2.332886219024658, "learning_rate": 1.8919283036551078e-05, - "loss": 0.2288, + "loss": 0.3238, "step": 3690 }, { "epoch": 1.7338331771321462, - "grad_norm": 1.6120171546936035, + "grad_norm": 1.7878814935684204, "learning_rate": 1.8916354264292408e-05, - "loss": 0.2122, + "loss": 0.2866, "step": 3700 }, { "epoch": 1.738519212746017, - "grad_norm": 1.8387460708618164, + "grad_norm": 2.455904722213745, "learning_rate": 1.891342549203374e-05, - "loss": 0.2292, + "loss": 0.2887, "step": 3710 }, { "epoch": 1.7432052483598874, - "grad_norm": 1.9124442338943481, + "grad_norm": 2.273310899734497, "learning_rate": 1.891049671977507e-05, - "loss": 0.2469, + "loss": 0.3377, "step": 3720 }, { "epoch": 1.7478912839737581, - "grad_norm": 2.267001152038574, + "grad_norm": 2.77280592918396, "learning_rate": 1.8907567947516403e-05, - "loss": 0.2139, + "loss": 0.2928, "step": 3730 }, { "epoch": 1.7525773195876289, - "grad_norm": 1.9495887756347656, + "grad_norm": 2.2917518615722656, "learning_rate": 1.8904639175257732e-05, - "loss": 0.2476, + "loss": 0.3154, "step": 3740 }, { "epoch": 1.7572633552014996, - "grad_norm": 1.8650192022323608, + "grad_norm": 3.0059680938720703, "learning_rate": 1.8901710402999065e-05, - "loss": 0.262, + "loss": 0.3216, "step": 3750 }, { "epoch": 1.7572633552014996, - "eval_loss": 0.04190748557448387, - "eval_pearson_cosine": 0.8188060652498521, - "eval_pearson_dot": 0.7306266271006336, - "eval_pearson_euclidean": 0.8067178845162717, - "eval_pearson_manhattan": 0.8079651581429825, - "eval_runtime": 3.6866, - "eval_samples_per_second": 406.882, - "eval_spearman_cosine": 0.8203796294288378, - "eval_spearman_dot": 0.7294293478148326, - "eval_spearman_euclidean": 0.8157749153170877, - "eval_spearman_manhattan": 0.8170299724489123, - "eval_steps_per_second": 25.498, + "eval_loss": 0.07067140191793442, + "eval_pearson_cosine": 0.7850573661327331, + "eval_pearson_dot": 0.6098455249231733, + "eval_pearson_euclidean": 0.7677167356750871, + "eval_pearson_manhattan": 0.7668202443827852, + "eval_runtime": 3.5184, + "eval_samples_per_second": 426.329, + "eval_spearman_cosine": 0.7949681180333369, + "eval_spearman_dot": 0.6135646781276878, + "eval_spearman_euclidean": 0.7791366533328872, + "eval_spearman_manhattan": 0.7786680488005274, + "eval_steps_per_second": 26.717, "step": 3750 }, { "epoch": 1.7619493908153703, - "grad_norm": 1.78036367893219, + "grad_norm": 2.2580602169036865, "learning_rate": 1.8898781630740395e-05, - "loss": 0.2441, + "loss": 0.2838, "step": 3760 }, { "epoch": 1.7666354264292408, - "grad_norm": 2.0995302200317383, + "grad_norm": 2.6639926433563232, "learning_rate": 1.8895852858481728e-05, - "loss": 0.2528, + "loss": 0.3388, "step": 3770 }, { "epoch": 1.7713214620431117, - "grad_norm": 1.7772722244262695, + "grad_norm": 2.4491138458251953, "learning_rate": 1.8892924086223057e-05, - "loss": 0.2305, + "loss": 0.2905, "step": 3780 }, { "epoch": 1.7760074976569822, - "grad_norm": 2.158984661102295, + "grad_norm": 2.4412620067596436, "learning_rate": 1.8889995313964387e-05, - "loss": 0.2374, + "loss": 0.2982, "step": 3790 }, { "epoch": 1.780693533270853, - "grad_norm": 1.888215184211731, + "grad_norm": 2.356416702270508, "learning_rate": 1.888706654170572e-05, - "loss": 0.2458, + "loss": 0.3302, "step": 3800 }, { "epoch": 1.7853795688847236, - "grad_norm": 2.109557867050171, + "grad_norm": 2.3429980278015137, "learning_rate": 1.888413776944705e-05, - "loss": 0.2463, + "loss": 0.3069, "step": 3810 }, { "epoch": 1.790065604498594, - "grad_norm": 1.715156078338623, + "grad_norm": 2.228518486022949, "learning_rate": 1.8881208997188382e-05, - "loss": 0.2421, + "loss": 0.3016, "step": 3820 }, { "epoch": 1.794751640112465, - "grad_norm": 2.860567808151245, + "grad_norm": 2.8037281036376953, "learning_rate": 1.887828022492971e-05, - "loss": 0.2728, + "loss": 0.3324, "step": 3830 }, { "epoch": 1.7994376757263355, - "grad_norm": 1.527709722518921, + "grad_norm": 1.893836498260498, "learning_rate": 1.887535145267104e-05, - "loss": 0.1982, + "loss": 0.2592, "step": 3840 }, { "epoch": 1.8041237113402062, - "grad_norm": 2.2306032180786133, + "grad_norm": 2.6253368854522705, "learning_rate": 1.8872422680412374e-05, - "loss": 0.2269, + "loss": 0.295, "step": 3850 }, { "epoch": 1.808809746954077, - "grad_norm": 1.285987377166748, + "grad_norm": 2.2410082817077637, "learning_rate": 1.8869493908153703e-05, - "loss": 0.1879, + "loss": 0.2495, "step": 3860 }, { "epoch": 1.8134957825679474, - "grad_norm": 1.7815241813659668, + "grad_norm": 2.3530426025390625, "learning_rate": 1.8866565135895033e-05, - "loss": 0.2337, + "loss": 0.3004, "step": 3870 }, { "epoch": 1.8181818181818183, - "grad_norm": 1.5507057905197144, + "grad_norm": 2.178670644760132, "learning_rate": 1.8863636363636366e-05, - "loss": 0.2542, + "loss": 0.3386, "step": 3880 }, { "epoch": 1.8228678537956888, - "grad_norm": 1.8540301322937012, + "grad_norm": 2.0660688877105713, "learning_rate": 1.8860707591377695e-05, - "loss": 0.2668, + "loss": 0.3117, "step": 3890 }, { "epoch": 1.8275538894095595, - "grad_norm": 1.5156047344207764, + "grad_norm": 2.5521416664123535, "learning_rate": 1.8857778819119025e-05, - "loss": 0.2438, + "loss": 0.3011, "step": 3900 }, { "epoch": 1.8322399250234302, - "grad_norm": 1.752375841140747, + "grad_norm": 2.3752758502960205, "learning_rate": 1.8854850046860358e-05, - "loss": 0.2324, + "loss": 0.3203, "step": 3910 }, { "epoch": 1.8369259606373007, - "grad_norm": 1.9103461503982544, + "grad_norm": 2.0836801528930664, "learning_rate": 1.8851921274601687e-05, - "loss": 0.2049, + "loss": 0.2706, "step": 3920 }, { "epoch": 1.8416119962511717, - "grad_norm": 2.043072462081909, + "grad_norm": 2.767982006072998, "learning_rate": 1.884899250234302e-05, - "loss": 0.2178, + "loss": 0.3077, "step": 3930 }, { "epoch": 1.8462980318650422, - "grad_norm": 1.7910971641540527, + "grad_norm": 2.7820098400115967, "learning_rate": 1.884606373008435e-05, - "loss": 0.2527, + "loss": 0.2991, "step": 3940 }, { "epoch": 1.8509840674789129, - "grad_norm": 2.146057605743408, + "grad_norm": 2.6888327598571777, "learning_rate": 1.8843134957825682e-05, - "loss": 0.2235, + "loss": 0.2862, "step": 3950 }, { "epoch": 1.8556701030927836, - "grad_norm": 2.3780410289764404, + "grad_norm": 2.4160315990448, "learning_rate": 1.8840206185567012e-05, - "loss": 0.2122, + "loss": 0.2725, "step": 3960 }, { "epoch": 1.860356138706654, - "grad_norm": 1.7693490982055664, + "grad_norm": 2.778198719024658, "learning_rate": 1.8837277413308345e-05, - "loss": 0.2663, + "loss": 0.3509, "step": 3970 }, { "epoch": 1.865042174320525, - "grad_norm": 1.9694123268127441, + "grad_norm": 1.8606864213943481, "learning_rate": 1.8834348641049674e-05, - "loss": 0.2357, + "loss": 0.2971, "step": 3980 }, { "epoch": 1.8697282099343955, - "grad_norm": 1.9589118957519531, + "grad_norm": 1.98170804977417, "learning_rate": 1.8831419868791004e-05, - "loss": 0.2549, + "loss": 0.3131, "step": 3990 }, { "epoch": 1.8744142455482662, - "grad_norm": 1.7447434663772583, + "grad_norm": 1.828123927116394, "learning_rate": 1.8828491096532337e-05, - "loss": 0.2269, + "loss": 0.3166, "step": 4000 }, { "epoch": 1.8744142455482662, - "eval_loss": 0.03927910327911377, - "eval_pearson_cosine": 0.8218360838151568, - "eval_pearson_dot": 0.7383648469379267, - "eval_pearson_euclidean": 0.79845453202784, - "eval_pearson_manhattan": 0.8001969992210543, - "eval_runtime": 3.608, - "eval_samples_per_second": 415.741, - "eval_spearman_cosine": 0.8235262235943273, - "eval_spearman_dot": 0.7374591899505454, - "eval_spearman_euclidean": 0.8094068736950194, - "eval_spearman_manhattan": 0.8111787861762638, - "eval_steps_per_second": 26.053, + "eval_loss": 0.07189416140317917, + "eval_pearson_cosine": 0.7798801652938181, + "eval_pearson_dot": 0.5736722449381091, + "eval_pearson_euclidean": 0.7563095863468678, + "eval_pearson_manhattan": 0.7550479478827015, + "eval_runtime": 3.2996, + "eval_samples_per_second": 454.607, + "eval_spearman_cosine": 0.7910837741143455, + "eval_spearman_dot": 0.5753555851409241, + "eval_spearman_euclidean": 0.7700926689739851, + "eval_spearman_manhattan": 0.7692694796240825, + "eval_steps_per_second": 28.489, "step": 4000 }, { "epoch": 1.879100281162137, - "grad_norm": 1.9940646886825562, + "grad_norm": 1.8813470602035522, "learning_rate": 1.8825562324273666e-05, - "loss": 0.2141, + "loss": 0.2484, "step": 4010 }, { "epoch": 1.8837863167760074, - "grad_norm": 1.5724692344665527, + "grad_norm": 2.005892038345337, "learning_rate": 1.8822633552014996e-05, - "loss": 0.2202, + "loss": 0.3137, "step": 4020 }, { "epoch": 1.8884723523898783, - "grad_norm": 1.880194067955017, + "grad_norm": 2.067732572555542, "learning_rate": 1.881970477975633e-05, - "loss": 0.245, + "loss": 0.3143, "step": 4030 }, { "epoch": 1.8931583880037488, - "grad_norm": 1.6564527750015259, + "grad_norm": 2.33829665184021, "learning_rate": 1.8816776007497658e-05, - "loss": 0.2367, + "loss": 0.3252, "step": 4040 }, { "epoch": 1.8978444236176195, - "grad_norm": 1.3603813648223877, + "grad_norm": 1.655529260635376, "learning_rate": 1.8813847235238987e-05, - "loss": 0.2219, + "loss": 0.2889, "step": 4050 }, { "epoch": 1.9025304592314902, - "grad_norm": 2.0585732460021973, + "grad_norm": 2.849501609802246, "learning_rate": 1.881091846298032e-05, - "loss": 0.27, + "loss": 0.323, "step": 4060 }, { "epoch": 1.9072164948453607, - "grad_norm": 2.4080166816711426, + "grad_norm": 2.8345351219177246, "learning_rate": 1.880798969072165e-05, - "loss": 0.2569, + "loss": 0.316, "step": 4070 }, { "epoch": 1.9119025304592316, - "grad_norm": 2.0260493755340576, + "grad_norm": 2.061378240585327, "learning_rate": 1.8805060918462983e-05, - "loss": 0.2239, + "loss": 0.3112, "step": 4080 }, { "epoch": 1.9165885660731021, - "grad_norm": 1.6909536123275757, + "grad_norm": 1.883039116859436, "learning_rate": 1.8802132146204312e-05, - "loss": 0.2229, + "loss": 0.2625, "step": 4090 }, { "epoch": 1.9212746016869728, - "grad_norm": 1.7024294137954712, + "grad_norm": 2.2609167098999023, "learning_rate": 1.8799203373945642e-05, - "loss": 0.2078, + "loss": 0.3043, "step": 4100 }, { "epoch": 1.9259606373008435, - "grad_norm": 1.7406731843948364, + "grad_norm": 2.27116060256958, "learning_rate": 1.8796274601686975e-05, - "loss": 0.2456, + "loss": 0.3022, "step": 4110 }, { "epoch": 1.930646672914714, - "grad_norm": 1.4349228143692017, + "grad_norm": 1.9829683303833008, "learning_rate": 1.8793345829428304e-05, - "loss": 0.2296, + "loss": 0.3147, "step": 4120 }, { "epoch": 1.935332708528585, - "grad_norm": 1.7777549028396606, + "grad_norm": 2.073610305786133, "learning_rate": 1.8790417057169637e-05, - "loss": 0.2121, + "loss": 0.2779, "step": 4130 }, { "epoch": 1.9400187441424555, - "grad_norm": 2.259178638458252, + "grad_norm": 2.183823585510254, "learning_rate": 1.8787488284910967e-05, - "loss": 0.2606, + "loss": 0.3041, "step": 4140 }, { "epoch": 1.9447047797563262, - "grad_norm": 2.10732364654541, + "grad_norm": 2.373502016067505, "learning_rate": 1.87845595126523e-05, - "loss": 0.2492, + "loss": 0.3037, "step": 4150 }, { "epoch": 1.9493908153701969, - "grad_norm": 2.2768290042877197, + "grad_norm": 3.190950393676758, "learning_rate": 1.878163074039363e-05, - "loss": 0.2155, + "loss": 0.2876, "step": 4160 }, { "epoch": 1.9540768509840674, - "grad_norm": 1.8602643013000488, + "grad_norm": 2.2904293537139893, "learning_rate": 1.8778701968134962e-05, - "loss": 0.2396, + "loss": 0.2957, "step": 4170 }, { "epoch": 1.9587628865979383, - "grad_norm": 2.197460412979126, + "grad_norm": 2.5620875358581543, "learning_rate": 1.877577319587629e-05, - "loss": 0.2169, + "loss": 0.2865, "step": 4180 }, { "epoch": 1.9634489222118088, - "grad_norm": 1.5060274600982666, + "grad_norm": 1.8882808685302734, "learning_rate": 1.877284442361762e-05, - "loss": 0.2306, + "loss": 0.302, "step": 4190 }, { "epoch": 1.9681349578256795, - "grad_norm": 2.116128921508789, + "grad_norm": 2.736853837966919, "learning_rate": 1.876991565135895e-05, - "loss": 0.2164, + "loss": 0.2783, "step": 4200 }, { "epoch": 1.9728209934395502, - "grad_norm": 2.097766399383545, + "grad_norm": 2.6126515865325928, "learning_rate": 1.8766986879100283e-05, - "loss": 0.2574, + "loss": 0.3337, "step": 4210 }, { "epoch": 1.9775070290534207, - "grad_norm": 1.824741244316101, + "grad_norm": 1.9608129262924194, "learning_rate": 1.8764058106841613e-05, - "loss": 0.249, + "loss": 0.2817, "step": 4220 }, { "epoch": 1.9821930646672916, - "grad_norm": 2.2704122066497803, + "grad_norm": 2.7429044246673584, "learning_rate": 1.8761129334582942e-05, - "loss": 0.2316, + "loss": 0.2827, "step": 4230 }, { "epoch": 1.986879100281162, - "grad_norm": 1.8285561800003052, + "grad_norm": 2.322411298751831, "learning_rate": 1.8758200562324275e-05, - "loss": 0.2489, + "loss": 0.2967, "step": 4240 }, { "epoch": 1.9915651358950328, - "grad_norm": 1.4909926652908325, + "grad_norm": 1.80522882938385, "learning_rate": 1.8755271790065605e-05, - "loss": 0.2472, + "loss": 0.315, "step": 4250 }, { "epoch": 1.9915651358950328, - "eval_loss": 0.03996381163597107, - "eval_pearson_cosine": 0.820344828031935, - "eval_pearson_dot": 0.7317039255282225, - "eval_pearson_euclidean": 0.8039609112737356, - "eval_pearson_manhattan": 0.8052691930531775, - "eval_runtime": 3.5603, - "eval_samples_per_second": 421.311, - "eval_spearman_cosine": 0.8224163636565568, - "eval_spearman_dot": 0.7308283663168708, - "eval_spearman_euclidean": 0.8146988265770505, - "eval_spearman_manhattan": 0.8160453182060199, - "eval_steps_per_second": 26.402, + "eval_loss": 0.0709637850522995, + "eval_pearson_cosine": 0.7818087092412611, + "eval_pearson_dot": 0.5917647670204929, + "eval_pearson_euclidean": 0.7671767812906438, + "eval_pearson_manhattan": 0.7657443892799449, + "eval_runtime": 3.3617, + "eval_samples_per_second": 446.198, + "eval_spearman_cosine": 0.7925261968709182, + "eval_spearman_dot": 0.5930341342679433, + "eval_spearman_euclidean": 0.7790026457642986, + "eval_spearman_manhattan": 0.7779600937996545, + "eval_steps_per_second": 27.962, "step": 4250 }, { "epoch": 1.9962511715089035, - "grad_norm": 1.7931636571884155, + "grad_norm": 2.3154516220092773, "learning_rate": 1.8752343017806937e-05, - "loss": 0.2324, + "loss": 0.3019, "step": 4260 }, { "epoch": 2.000937207122774, - "grad_norm": 1.267348289489746, + "grad_norm": 1.769853949546814, "learning_rate": 1.8749414245548267e-05, - "loss": 0.2053, + "loss": 0.2639, "step": 4270 }, { "epoch": 2.005623242736645, - "grad_norm": 1.977459192276001, + "grad_norm": 2.7089197635650635, "learning_rate": 1.87464854732896e-05, - "loss": 0.1835, + "loss": 0.1872, "step": 4280 }, { "epoch": 2.0103092783505154, - "grad_norm": 1.8532277345657349, + "grad_norm": 2.1212291717529297, "learning_rate": 1.874355670103093e-05, - "loss": 0.1941, + "loss": 0.189, "step": 4290 }, { "epoch": 2.014995313964386, - "grad_norm": 1.478722095489502, + "grad_norm": 1.8745218515396118, "learning_rate": 1.874062792877226e-05, - "loss": 0.1674, + "loss": 0.1749, "step": 4300 }, { "epoch": 2.019681349578257, - "grad_norm": 1.7603696584701538, + "grad_norm": 2.0145514011383057, "learning_rate": 1.8737699156513592e-05, - "loss": 0.1752, + "loss": 0.1809, "step": 4310 }, { "epoch": 2.0243673851921273, - "grad_norm": 2.0662457942962646, + "grad_norm": 2.370055913925171, "learning_rate": 1.873477038425492e-05, - "loss": 0.1638, + "loss": 0.2056, "step": 4320 }, { "epoch": 2.0290534208059983, - "grad_norm": 1.3146560192108154, + "grad_norm": 1.7560367584228516, "learning_rate": 1.8731841611996254e-05, - "loss": 0.1368, + "loss": 0.1742, "step": 4330 }, { "epoch": 2.0337394564198688, - "grad_norm": 1.8261940479278564, + "grad_norm": 2.693603277206421, "learning_rate": 1.8728912839737584e-05, - "loss": 0.1643, + "loss": 0.2017, "step": 4340 }, { "epoch": 2.0384254920337392, - "grad_norm": 1.4603139162063599, + "grad_norm": 2.1456475257873535, "learning_rate": 1.8725984067478917e-05, - "loss": 0.1655, + "loss": 0.1967, "step": 4350 }, { "epoch": 2.04311152764761, - "grad_norm": 1.811943531036377, + "grad_norm": 1.8832343816757202, "learning_rate": 1.8723055295220246e-05, - "loss": 0.1617, + "loss": 0.182, "step": 4360 }, { "epoch": 2.0477975632614807, - "grad_norm": 2.325777769088745, + "grad_norm": 2.2446205615997314, "learning_rate": 1.8720126522961576e-05, - "loss": 0.1708, + "loss": 0.1895, "step": 4370 }, { "epoch": 2.0524835988753516, - "grad_norm": 1.7482625246047974, + "grad_norm": 1.8662675619125366, "learning_rate": 1.871719775070291e-05, - "loss": 0.1758, + "loss": 0.2091, "step": 4380 }, { "epoch": 2.057169634489222, - "grad_norm": 1.916061520576477, + "grad_norm": 2.3427059650421143, "learning_rate": 1.8714268978444238e-05, - "loss": 0.1729, + "loss": 0.2046, "step": 4390 }, { "epoch": 2.0618556701030926, - "grad_norm": 1.4515407085418701, + "grad_norm": 1.8921583890914917, "learning_rate": 1.8711340206185567e-05, - "loss": 0.1739, + "loss": 0.21, "step": 4400 }, { "epoch": 2.0665417057169635, - "grad_norm": 1.7637025117874146, + "grad_norm": 2.3345112800598145, "learning_rate": 1.8708411433926897e-05, - "loss": 0.1648, + "loss": 0.211, "step": 4410 }, { "epoch": 2.071227741330834, - "grad_norm": 2.1204230785369873, + "grad_norm": 2.8094024658203125, "learning_rate": 1.870548266166823e-05, - "loss": 0.1725, + "loss": 0.185, "step": 4420 }, { "epoch": 2.075913776944705, - "grad_norm": 1.7462095022201538, + "grad_norm": 2.182375192642212, "learning_rate": 1.870255388940956e-05, - "loss": 0.154, + "loss": 0.1743, "step": 4430 }, { "epoch": 2.0805998125585754, - "grad_norm": 1.6449689865112305, + "grad_norm": 1.9309943914413452, "learning_rate": 1.8699625117150892e-05, - "loss": 0.1651, + "loss": 0.2002, "step": 4440 }, { "epoch": 2.085285848172446, - "grad_norm": 1.8776732683181763, + "grad_norm": 2.410891532897949, "learning_rate": 1.869669634489222e-05, - "loss": 0.1752, + "loss": 0.2034, "step": 4450 }, { "epoch": 2.089971883786317, - "grad_norm": 1.5373666286468506, + "grad_norm": 1.9030256271362305, "learning_rate": 1.8693767572633555e-05, - "loss": 0.1551, + "loss": 0.1985, "step": 4460 }, { "epoch": 2.0946579194001873, - "grad_norm": 1.9183931350708008, + "grad_norm": 2.387491226196289, "learning_rate": 1.8690838800374884e-05, - "loss": 0.1562, + "loss": 0.1657, "step": 4470 }, { "epoch": 2.0993439550140582, - "grad_norm": 2.184581995010376, + "grad_norm": 2.448557138442993, "learning_rate": 1.8687910028116217e-05, - "loss": 0.1544, + "loss": 0.1889, "step": 4480 }, { "epoch": 2.1040299906279287, - "grad_norm": 1.1829646825790405, + "grad_norm": 1.8442336320877075, "learning_rate": 1.8684981255857546e-05, - "loss": 0.1474, + "loss": 0.1638, "step": 4490 }, { "epoch": 2.108716026241799, - "grad_norm": 1.7661231756210327, + "grad_norm": 2.192814588546753, "learning_rate": 1.8682052483598876e-05, - "loss": 0.1838, + "loss": 0.2117, "step": 4500 }, { "epoch": 2.108716026241799, - "eval_loss": 0.03478589281439781, - "eval_pearson_cosine": 0.8183668280938576, - "eval_pearson_dot": 0.7494542878437821, - "eval_pearson_euclidean": 0.8004851343617361, - "eval_pearson_manhattan": 0.8022690605878324, - "eval_runtime": 3.3337, - "eval_samples_per_second": 449.946, - "eval_spearman_cosine": 0.819061603084573, - "eval_spearman_dot": 0.7480798887745068, - "eval_spearman_euclidean": 0.8085010936446739, - "eval_spearman_manhattan": 0.8099334586781558, - "eval_steps_per_second": 28.197, + "eval_loss": 0.0544867068529129, + "eval_pearson_cosine": 0.7772149225219422, + "eval_pearson_dot": 0.6058722441499071, + "eval_pearson_euclidean": 0.7566861769282127, + "eval_pearson_manhattan": 0.7550653692259459, + "eval_runtime": 3.562, + "eval_samples_per_second": 421.112, + "eval_spearman_cosine": 0.7889882574404774, + "eval_spearman_dot": 0.609585139530218, + "eval_spearman_euclidean": 0.7711664957300913, + "eval_spearman_manhattan": 0.7702480077346849, + "eval_steps_per_second": 26.39, "step": 4500 }, { "epoch": 2.11340206185567, - "grad_norm": 1.5288666486740112, + "grad_norm": 1.964051604270935, "learning_rate": 1.867912371134021e-05, - "loss": 0.1483, + "loss": 0.1727, "step": 4510 }, { "epoch": 2.1180880974695406, - "grad_norm": 1.8518108129501343, + "grad_norm": 2.180359363555908, "learning_rate": 1.867619493908154e-05, - "loss": 0.1556, + "loss": 0.1929, "step": 4520 }, { "epoch": 2.1227741330834116, - "grad_norm": 1.7130534648895264, + "grad_norm": 1.5920389890670776, "learning_rate": 1.867326616682287e-05, - "loss": 0.2054, + "loss": 0.215, "step": 4530 }, { "epoch": 2.127460168697282, - "grad_norm": 1.6781951189041138, + "grad_norm": 2.0330374240875244, "learning_rate": 1.86703373945642e-05, - "loss": 0.1563, + "loss": 0.1782, "step": 4540 }, { "epoch": 2.1321462043111525, - "grad_norm": 1.960902452468872, + "grad_norm": 3.1396799087524414, "learning_rate": 1.866740862230553e-05, - "loss": 0.1666, + "loss": 0.1843, "step": 4550 }, { "epoch": 2.1368322399250235, - "grad_norm": 1.9517632722854614, + "grad_norm": 2.507819652557373, "learning_rate": 1.8664479850046863e-05, - "loss": 0.1863, + "loss": 0.2272, "step": 4560 }, { "epoch": 2.141518275538894, - "grad_norm": 1.7729823589324951, + "grad_norm": 1.9377321004867554, "learning_rate": 1.8661551077788193e-05, - "loss": 0.1783, + "loss": 0.2003, "step": 4570 }, { "epoch": 2.146204311152765, - "grad_norm": 2.2520790100097656, + "grad_norm": 2.868741512298584, "learning_rate": 1.8658622305529522e-05, - "loss": 0.208, + "loss": 0.2266, "step": 4580 }, { "epoch": 2.1508903467666354, - "grad_norm": 1.4760936498641968, + "grad_norm": 1.8727847337722778, "learning_rate": 1.8655693533270855e-05, - "loss": 0.1842, + "loss": 0.1965, "step": 4590 }, { "epoch": 2.155576382380506, - "grad_norm": 1.8456919193267822, + "grad_norm": 1.937414526939392, "learning_rate": 1.8652764761012184e-05, - "loss": 0.1616, + "loss": 0.1957, "step": 4600 }, { "epoch": 2.160262417994377, - "grad_norm": 1.528846263885498, + "grad_norm": 2.084174156188965, "learning_rate": 1.8649835988753514e-05, - "loss": 0.1342, + "loss": 0.1668, "step": 4610 }, { "epoch": 2.1649484536082473, - "grad_norm": 1.1334350109100342, + "grad_norm": 1.846552848815918, "learning_rate": 1.8646907216494847e-05, - "loss": 0.1655, + "loss": 0.184, "step": 4620 }, { "epoch": 2.169634489222118, - "grad_norm": 1.359473705291748, + "grad_norm": 1.9087766408920288, "learning_rate": 1.8643978444236176e-05, - "loss": 0.1667, + "loss": 0.1961, "step": 4630 }, { "epoch": 2.1743205248359887, - "grad_norm": 1.6595484018325806, + "grad_norm": 2.011108875274658, "learning_rate": 1.864104967197751e-05, - "loss": 0.1597, + "loss": 0.1848, "step": 4640 }, { "epoch": 2.179006560449859, - "grad_norm": 1.9534510374069214, + "grad_norm": 2.8976380825042725, "learning_rate": 1.863812089971884e-05, - "loss": 0.18, + "loss": 0.2082, "step": 4650 }, { "epoch": 2.18369259606373, - "grad_norm": 2.137612819671631, + "grad_norm": 2.1731081008911133, "learning_rate": 1.863519212746017e-05, - "loss": 0.1649, + "loss": 0.1852, "step": 4660 }, { "epoch": 2.1883786316776006, - "grad_norm": 1.4351214170455933, + "grad_norm": 1.7672990560531616, "learning_rate": 1.86322633552015e-05, - "loss": 0.1798, + "loss": 0.2023, "step": 4670 }, { "epoch": 2.1930646672914715, - "grad_norm": 1.5940383672714233, + "grad_norm": 1.6886131763458252, "learning_rate": 1.8629334582942834e-05, - "loss": 0.1762, + "loss": 0.2126, "step": 4680 }, { "epoch": 2.197750702905342, - "grad_norm": 1.1390234231948853, + "grad_norm": 1.4913769960403442, "learning_rate": 1.8626405810684164e-05, - "loss": 0.1575, + "loss": 0.1836, "step": 4690 }, { "epoch": 2.2024367385192125, - "grad_norm": 1.7934620380401611, + "grad_norm": 2.0632176399230957, "learning_rate": 1.8623477038425493e-05, - "loss": 0.16, + "loss": 0.2048, "step": 4700 }, { "epoch": 2.2071227741330834, - "grad_norm": 1.5120437145233154, + "grad_norm": 1.9084508419036865, "learning_rate": 1.8620548266166826e-05, - "loss": 0.1711, + "loss": 0.1836, "step": 4710 }, { "epoch": 2.211808809746954, - "grad_norm": 1.8935743570327759, + "grad_norm": 1.9352916479110718, "learning_rate": 1.8617619493908155e-05, - "loss": 0.1396, + "loss": 0.1653, "step": 4720 }, { "epoch": 2.216494845360825, - "grad_norm": 1.7621512413024902, + "grad_norm": 1.974461317062378, "learning_rate": 1.8614690721649485e-05, - "loss": 0.1953, + "loss": 0.2053, "step": 4730 }, { "epoch": 2.2211808809746953, - "grad_norm": 9.526521682739258, + "grad_norm": 2.1307456493377686, "learning_rate": 1.8611761949390818e-05, - "loss": 0.1781, + "loss": 0.1835, "step": 4740 }, { "epoch": 2.2258669165885663, - "grad_norm": 1.4335854053497314, + "grad_norm": 1.413037896156311, "learning_rate": 1.8608833177132147e-05, - "loss": 0.1509, + "loss": 0.1725, "step": 4750 }, { "epoch": 2.2258669165885663, - "eval_loss": 0.03594741225242615, - "eval_pearson_cosine": 0.8116579897427059, - "eval_pearson_dot": 0.734380652885136, - "eval_pearson_euclidean": 0.7957971930124614, - "eval_pearson_manhattan": 0.7976626643154123, - "eval_runtime": 3.4015, - "eval_samples_per_second": 440.981, - "eval_spearman_cosine": 0.8119603885409888, - "eval_spearman_dot": 0.7343109043661656, - "eval_spearman_euclidean": 0.803610294766424, - "eval_spearman_manhattan": 0.8054368058373911, - "eval_steps_per_second": 27.635, + "eval_loss": 0.05439988151192665, + "eval_pearson_cosine": 0.7779719017710534, + "eval_pearson_dot": 0.6064920766481805, + "eval_pearson_euclidean": 0.7604967084059453, + "eval_pearson_manhattan": 0.7592805758538006, + "eval_runtime": 3.2739, + "eval_samples_per_second": 458.175, + "eval_spearman_cosine": 0.786759628149343, + "eval_spearman_dot": 0.6127818290418755, + "eval_spearman_euclidean": 0.7721037448944629, + "eval_spearman_manhattan": 0.7714373415381524, + "eval_steps_per_second": 28.712, "step": 4750 }, { "epoch": 2.2305529522024368, - "grad_norm": 1.9357318878173828, + "grad_norm": 2.033710241317749, "learning_rate": 1.8605904404873477e-05, - "loss": 0.1464, + "loss": 0.1769, "step": 4760 }, { "epoch": 2.2352389878163073, - "grad_norm": 1.958662986755371, + "grad_norm": 2.065793752670288, "learning_rate": 1.860297563261481e-05, - "loss": 0.1699, + "loss": 0.1896, "step": 4770 }, { "epoch": 2.239925023430178, - "grad_norm": 1.9908474683761597, + "grad_norm": 2.0393176078796387, "learning_rate": 1.860004686035614e-05, - "loss": 0.1763, + "loss": 0.1944, "step": 4780 }, { "epoch": 2.2446110590440487, - "grad_norm": 1.8446073532104492, + "grad_norm": 1.967092514038086, "learning_rate": 1.859711808809747e-05, - "loss": 0.1972, + "loss": 0.218, "step": 4790 }, { "epoch": 2.2492970946579196, - "grad_norm": 1.9557933807373047, + "grad_norm": 2.614811420440674, "learning_rate": 1.85941893158388e-05, - "loss": 0.177, + "loss": 0.2158, "step": 4800 }, { "epoch": 2.25398313027179, - "grad_norm": 1.9075273275375366, + "grad_norm": 1.9915626049041748, "learning_rate": 1.859126054358013e-05, - "loss": 0.1719, + "loss": 0.199, "step": 4810 }, { "epoch": 2.2586691658856606, - "grad_norm": 1.6813454627990723, + "grad_norm": 2.308933973312378, "learning_rate": 1.8588331771321464e-05, - "loss": 0.1555, + "loss": 0.1746, "step": 4820 }, { "epoch": 2.2633552014995315, - "grad_norm": 1.677162528038025, + "grad_norm": 2.2984583377838135, "learning_rate": 1.8585402999062793e-05, - "loss": 0.1585, + "loss": 0.1905, "step": 4830 }, { "epoch": 2.268041237113402, - "grad_norm": 1.9858368635177612, + "grad_norm": 2.0273101329803467, "learning_rate": 1.8582474226804126e-05, - "loss": 0.1627, + "loss": 0.1655, "step": 4840 }, { "epoch": 2.2727272727272725, - "grad_norm": 1.6848777532577515, + "grad_norm": 1.745671272277832, "learning_rate": 1.8579545454545456e-05, - "loss": 0.1576, + "loss": 0.1745, "step": 4850 }, { "epoch": 2.2774133083411434, - "grad_norm": 2.0122177600860596, + "grad_norm": 1.9413400888442993, "learning_rate": 1.857661668228679e-05, - "loss": 0.172, + "loss": 0.1899, "step": 4860 }, { "epoch": 2.282099343955014, - "grad_norm": 1.666049599647522, + "grad_norm": 2.113260269165039, "learning_rate": 1.8573687910028118e-05, - "loss": 0.1554, + "loss": 0.1826, "step": 4870 }, { "epoch": 2.286785379568885, - "grad_norm": 1.4426305294036865, + "grad_norm": 1.9468377828598022, "learning_rate": 1.857075913776945e-05, - "loss": 0.1268, + "loss": 0.1596, "step": 4880 }, { "epoch": 2.2914714151827553, - "grad_norm": 1.8807013034820557, + "grad_norm": 2.1250803470611572, "learning_rate": 1.856783036551078e-05, - "loss": 0.1892, + "loss": 0.1969, "step": 4890 }, { "epoch": 2.296157450796626, - "grad_norm": 1.6185581684112549, + "grad_norm": 1.5394418239593506, "learning_rate": 1.856490159325211e-05, - "loss": 0.1645, + "loss": 0.1731, "step": 4900 }, { "epoch": 2.3008434864104967, - "grad_norm": 1.7439680099487305, + "grad_norm": 2.2445003986358643, "learning_rate": 1.8561972820993443e-05, - "loss": 0.1677, + "loss": 0.1817, "step": 4910 }, { "epoch": 2.3055295220243672, - "grad_norm": 1.9492429494857788, + "grad_norm": 2.201733350753784, "learning_rate": 1.8559044048734773e-05, - "loss": 0.1626, + "loss": 0.175, "step": 4920 }, { "epoch": 2.310215557638238, - "grad_norm": 1.7018874883651733, + "grad_norm": 2.077092409133911, "learning_rate": 1.8556115276476102e-05, - "loss": 0.1581, + "loss": 0.1755, "step": 4930 }, { "epoch": 2.3149015932521086, - "grad_norm": 1.5858055353164673, + "grad_norm": 2.0630977153778076, "learning_rate": 1.855318650421743e-05, - "loss": 0.2029, + "loss": 0.1967, "step": 4940 }, { "epoch": 2.319587628865979, - "grad_norm": 2.1818623542785645, + "grad_norm": 2.737861156463623, "learning_rate": 1.8550257731958764e-05, - "loss": 0.2094, + "loss": 0.2175, "step": 4950 }, { "epoch": 2.32427366447985, - "grad_norm": 1.5739350318908691, + "grad_norm": 2.0301871299743652, "learning_rate": 1.8547328959700094e-05, - "loss": 0.1565, + "loss": 0.1938, "step": 4960 }, { "epoch": 2.3289597000937206, - "grad_norm": 1.1287211179733276, + "grad_norm": 1.6436868906021118, "learning_rate": 1.8544400187441427e-05, - "loss": 0.1982, + "loss": 0.2215, "step": 4970 }, { "epoch": 2.3336457357075915, - "grad_norm": 1.8876228332519531, + "grad_norm": 2.0046422481536865, "learning_rate": 1.8541471415182756e-05, - "loss": 0.1571, + "loss": 0.189, "step": 4980 }, { "epoch": 2.338331771321462, - "grad_norm": 2.1654696464538574, + "grad_norm": 2.598639488220215, "learning_rate": 1.8538542642924086e-05, - "loss": 0.163, + "loss": 0.2075, "step": 4990 }, { "epoch": 2.3430178069353325, - "grad_norm": 1.3975324630737305, + "grad_norm": 1.8523390293121338, "learning_rate": 1.853561387066542e-05, - "loss": 0.1816, + "loss": 0.1985, "step": 5000 }, { "epoch": 2.3430178069353325, - "eval_loss": 0.033037662506103516, - "eval_pearson_cosine": 0.8184973777993747, - "eval_pearson_dot": 0.750740290890306, - "eval_pearson_euclidean": 0.7978417286720072, - "eval_pearson_manhattan": 0.7999562879189881, - "eval_runtime": 3.3021, - "eval_samples_per_second": 454.257, - "eval_spearman_cosine": 0.8181019655563226, - "eval_spearman_dot": 0.750101463474286, - "eval_spearman_euclidean": 0.8059752038134308, - "eval_spearman_manhattan": 0.807850996089844, - "eval_steps_per_second": 28.467, + "eval_loss": 0.05401456356048584, + "eval_pearson_cosine": 0.781767118943602, + "eval_pearson_dot": 0.6017329721420595, + "eval_pearson_euclidean": 0.7626066877478763, + "eval_pearson_manhattan": 0.7620550203728129, + "eval_runtime": 3.2064, + "eval_samples_per_second": 467.808, + "eval_spearman_cosine": 0.7915644905642973, + "eval_spearman_dot": 0.6077979618715047, + "eval_spearman_euclidean": 0.773425356513352, + "eval_spearman_manhattan": 0.7732819218811388, + "eval_steps_per_second": 29.316, "step": 5000 }, { "epoch": 2.3477038425492034, - "grad_norm": 1.9489047527313232, + "grad_norm": 2.0773608684539795, "learning_rate": 1.8532685098406748e-05, - "loss": 0.1478, + "loss": 0.1749, "step": 5010 }, { "epoch": 2.352389878163074, - "grad_norm": 1.7943733930587769, + "grad_norm": 1.6482021808624268, "learning_rate": 1.852975632614808e-05, - "loss": 0.1447, + "loss": 0.1719, "step": 5020 }, { "epoch": 2.357075913776945, - "grad_norm": 1.9680578708648682, + "grad_norm": 2.1038990020751953, "learning_rate": 1.852682755388941e-05, - "loss": 0.158, + "loss": 0.2014, "step": 5030 }, { "epoch": 2.3617619493908153, - "grad_norm": 1.76882004737854, + "grad_norm": 2.396871566772461, "learning_rate": 1.8523898781630743e-05, - "loss": 0.1616, + "loss": 0.1855, "step": 5040 }, { "epoch": 2.3664479850046862, - "grad_norm": 1.7545366287231445, + "grad_norm": 2.3098223209381104, "learning_rate": 1.8520970009372073e-05, - "loss": 0.1748, + "loss": 0.1819, "step": 5050 }, { "epoch": 2.3711340206185567, - "grad_norm": 1.8479957580566406, + "grad_norm": 1.9691858291625977, "learning_rate": 1.8518041237113406e-05, - "loss": 0.1536, + "loss": 0.1786, "step": 5060 }, { "epoch": 2.375820056232427, - "grad_norm": 2.2647321224212646, + "grad_norm": 2.8055005073547363, "learning_rate": 1.8515112464854735e-05, - "loss": 0.169, + "loss": 0.1864, "step": 5070 }, { "epoch": 2.380506091846298, - "grad_norm": 2.5361509323120117, + "grad_norm": 2.2753820419311523, "learning_rate": 1.8512183692596065e-05, - "loss": 0.1664, + "loss": 0.1918, "step": 5080 }, { "epoch": 2.3851921274601686, - "grad_norm": 1.6961290836334229, + "grad_norm": 1.632279634475708, "learning_rate": 1.8509254920337398e-05, - "loss": 0.1581, + "loss": 0.1767, "step": 5090 }, { "epoch": 2.3898781630740396, - "grad_norm": 2.0420143604278564, + "grad_norm": 2.177457571029663, "learning_rate": 1.8506326148078727e-05, - "loss": 0.1761, + "loss": 0.1866, "step": 5100 }, { "epoch": 2.39456419868791, - "grad_norm": 1.5825088024139404, + "grad_norm": 1.8641494512557983, "learning_rate": 1.8503397375820057e-05, - "loss": 0.1462, + "loss": 0.1833, "step": 5110 }, { "epoch": 2.3992502343017805, - "grad_norm": 1.1550185680389404, + "grad_norm": 2.0116262435913086, "learning_rate": 1.8500468603561386e-05, - "loss": 0.1612, + "loss": 0.1975, "step": 5120 }, { "epoch": 2.4039362699156515, - "grad_norm": 1.1477668285369873, + "grad_norm": 1.130846381187439, "learning_rate": 1.849753983130272e-05, - "loss": 0.168, + "loss": 0.1992, "step": 5130 }, { "epoch": 2.408622305529522, - "grad_norm": 2.5751688480377197, + "grad_norm": 2.743990421295166, "learning_rate": 1.849461105904405e-05, - "loss": 0.1737, + "loss": 0.1996, "step": 5140 }, { "epoch": 2.413308341143393, - "grad_norm": 0.8213298916816711, + "grad_norm": 2.4368762969970703, "learning_rate": 1.849168228678538e-05, - "loss": 0.1528, + "loss": 0.1722, "step": 5150 }, { "epoch": 2.4179943767572634, - "grad_norm": 2.1841747760772705, + "grad_norm": 2.659679889678955, "learning_rate": 1.848875351452671e-05, - "loss": 0.1658, + "loss": 0.1834, "step": 5160 }, { "epoch": 2.422680412371134, - "grad_norm": 1.1756603717803955, + "grad_norm": 1.9601231813430786, "learning_rate": 1.8485824742268044e-05, - "loss": 0.1593, + "loss": 0.1943, "step": 5170 }, { "epoch": 2.427366447985005, - "grad_norm": 1.4406476020812988, + "grad_norm": 1.521337628364563, "learning_rate": 1.8482895970009373e-05, - "loss": 0.1725, + "loss": 0.2009, "step": 5180 }, { "epoch": 2.4320524835988753, - "grad_norm": 1.0863920450210571, + "grad_norm": 1.4766628742218018, "learning_rate": 1.8479967197750703e-05, - "loss": 0.1629, + "loss": 0.1848, "step": 5190 }, { "epoch": 2.436738519212746, - "grad_norm": 1.736379861831665, + "grad_norm": 2.3502743244171143, "learning_rate": 1.8477038425492036e-05, - "loss": 0.1452, + "loss": 0.165, "step": 5200 }, { "epoch": 2.4414245548266167, - "grad_norm": 1.3213779926300049, + "grad_norm": 1.5100184679031372, "learning_rate": 1.8474109653233365e-05, - "loss": 0.1654, + "loss": 0.1983, "step": 5210 }, { "epoch": 2.446110590440487, - "grad_norm": 1.8341721296310425, + "grad_norm": 2.336106061935425, "learning_rate": 1.8471180880974698e-05, - "loss": 0.149, + "loss": 0.1871, "step": 5220 }, { "epoch": 2.450796626054358, - "grad_norm": 1.6373144388198853, + "grad_norm": 2.1695876121520996, "learning_rate": 1.8468252108716028e-05, - "loss": 0.1769, + "loss": 0.1996, "step": 5230 }, { "epoch": 2.4554826616682286, - "grad_norm": 1.9199351072311401, + "grad_norm": 2.047522783279419, "learning_rate": 1.846532333645736e-05, - "loss": 0.1848, + "loss": 0.2088, "step": 5240 }, { "epoch": 2.4601686972820995, - "grad_norm": 1.5321757793426514, + "grad_norm": 1.9469248056411743, "learning_rate": 1.846239456419869e-05, - "loss": 0.166, + "loss": 0.1871, "step": 5250 }, { "epoch": 2.4601686972820995, - "eval_loss": 0.033459678292274475, - "eval_pearson_cosine": 0.81833954439713, - "eval_pearson_dot": 0.7450131051603677, - "eval_pearson_euclidean": 0.7997398913396002, - "eval_pearson_manhattan": 0.8015232151946456, - "eval_runtime": 3.3372, - "eval_samples_per_second": 449.478, - "eval_spearman_cosine": 0.818760866631701, - "eval_spearman_dot": 0.7445336159751238, - "eval_spearman_euclidean": 0.8090651280572486, - "eval_spearman_manhattan": 0.8107108068343578, - "eval_steps_per_second": 28.167, + "eval_loss": 0.05267513543367386, + "eval_pearson_cosine": 0.7830377509621513, + "eval_pearson_dot": 0.5843463467596735, + "eval_pearson_euclidean": 0.7586592990276699, + "eval_pearson_manhattan": 0.7576312827364555, + "eval_runtime": 3.3095, + "eval_samples_per_second": 453.24, + "eval_spearman_cosine": 0.7898144287463136, + "eval_spearman_dot": 0.5893690662486756, + "eval_spearman_euclidean": 0.7723820863953171, + "eval_spearman_manhattan": 0.7718409919252442, + "eval_steps_per_second": 28.403, "step": 5250 }, { "epoch": 2.46485473289597, - "grad_norm": 2.2415823936462402, + "grad_norm": 2.1970055103302, "learning_rate": 1.845946579194002e-05, - "loss": 0.1606, + "loss": 0.2059, "step": 5260 }, { "epoch": 2.4695407685098405, - "grad_norm": 0.9711971879005432, + "grad_norm": 1.7079726457595825, "learning_rate": 1.8456537019681352e-05, - "loss": 0.1546, + "loss": 0.1677, "step": 5270 }, { "epoch": 2.4742268041237114, - "grad_norm": 1.7712616920471191, + "grad_norm": 2.1129045486450195, "learning_rate": 1.8453608247422682e-05, - "loss": 0.1725, + "loss": 0.1774, "step": 5280 }, { "epoch": 2.478912839737582, - "grad_norm": 1.9109606742858887, + "grad_norm": 2.0570807456970215, "learning_rate": 1.845067947516401e-05, - "loss": 0.175, + "loss": 0.1863, "step": 5290 }, { "epoch": 2.483598875351453, - "grad_norm": 1.4726054668426514, + "grad_norm": 2.095052719116211, "learning_rate": 1.8447750702905344e-05, - "loss": 0.1432, + "loss": 0.1675, "step": 5300 }, { "epoch": 2.4882849109653233, - "grad_norm": 1.5022660493850708, + "grad_norm": 2.5408618450164795, "learning_rate": 1.8444821930646674e-05, - "loss": 0.1682, + "loss": 0.2033, "step": 5310 }, { "epoch": 2.492970946579194, - "grad_norm": 2.4403464794158936, + "grad_norm": 2.0885772705078125, "learning_rate": 1.8441893158388003e-05, - "loss": 0.1554, + "loss": 0.178, "step": 5320 }, { "epoch": 2.4976569821930648, - "grad_norm": 2.6438190937042236, + "grad_norm": 2.907564640045166, "learning_rate": 1.8438964386129336e-05, - "loss": 0.1874, + "loss": 0.2121, "step": 5330 }, { "epoch": 2.5023430178069352, - "grad_norm": 1.2486696243286133, + "grad_norm": 1.3953640460968018, "learning_rate": 1.8436035613870666e-05, - "loss": 0.179, + "loss": 0.1838, "step": 5340 }, { "epoch": 2.5070290534208057, - "grad_norm": 1.551320195198059, + "grad_norm": 1.825296401977539, "learning_rate": 1.8433106841612e-05, - "loss": 0.1498, + "loss": 0.1688, "step": 5350 }, { "epoch": 2.5117150890346767, - "grad_norm": 1.7184678316116333, + "grad_norm": 2.5563881397247314, "learning_rate": 1.8430178069353328e-05, - "loss": 0.1643, + "loss": 0.1798, "step": 5360 }, { "epoch": 2.5164011246485476, - "grad_norm": 1.7342448234558105, + "grad_norm": 2.355055570602417, "learning_rate": 1.842724929709466e-05, - "loss": 0.158, + "loss": 0.1858, "step": 5370 }, { "epoch": 2.521087160262418, - "grad_norm": 1.900806188583374, + "grad_norm": 2.4546165466308594, "learning_rate": 1.842432052483599e-05, - "loss": 0.1706, + "loss": 0.2149, "step": 5380 }, { "epoch": 2.5257731958762886, - "grad_norm": 1.599795937538147, + "grad_norm": 2.0621817111968994, "learning_rate": 1.842139175257732e-05, - "loss": 0.1878, + "loss": 0.2215, "step": 5390 }, { "epoch": 2.530459231490159, - "grad_norm": 1.4454323053359985, + "grad_norm": 1.7873655557632446, "learning_rate": 1.8418462980318653e-05, - "loss": 0.1655, + "loss": 0.1872, "step": 5400 }, { "epoch": 2.53514526710403, - "grad_norm": 1.7434407472610474, + "grad_norm": 1.8408267498016357, "learning_rate": 1.8415534208059982e-05, - "loss": 0.1503, + "loss": 0.181, "step": 5410 }, { "epoch": 2.539831302717901, - "grad_norm": 2.328422784805298, + "grad_norm": 2.56502628326416, "learning_rate": 1.8412605435801315e-05, - "loss": 0.1702, + "loss": 0.1946, "step": 5420 }, { "epoch": 2.5445173383317714, - "grad_norm": 1.7260003089904785, + "grad_norm": 1.9909722805023193, "learning_rate": 1.8409676663542645e-05, - "loss": 0.1624, + "loss": 0.1912, "step": 5430 }, { "epoch": 2.549203373945642, - "grad_norm": 1.7711926698684692, + "grad_norm": 2.530014753341675, "learning_rate": 1.8406747891283974e-05, - "loss": 0.1609, + "loss": 0.1898, "step": 5440 }, { "epoch": 2.5538894095595124, - "grad_norm": 1.963535189628601, + "grad_norm": 2.452223300933838, "learning_rate": 1.8403819119025307e-05, - "loss": 0.1881, + "loss": 0.2114, "step": 5450 }, { "epoch": 2.5585754451733833, - "grad_norm": 1.5332142114639282, + "grad_norm": 2.0805137157440186, "learning_rate": 1.8400890346766637e-05, - "loss": 0.1891, + "loss": 0.2237, "step": 5460 }, { "epoch": 2.5632614807872542, - "grad_norm": 1.4771735668182373, + "grad_norm": 1.8469469547271729, "learning_rate": 1.8397961574507966e-05, - "loss": 0.1455, + "loss": 0.1895, "step": 5470 }, { "epoch": 2.5679475164011247, - "grad_norm": 1.3587532043457031, + "grad_norm": 2.0643043518066406, "learning_rate": 1.83950328022493e-05, - "loss": 0.1579, + "loss": 0.1943, "step": 5480 }, { "epoch": 2.572633552014995, - "grad_norm": 1.5083823204040527, + "grad_norm": 2.1368064880371094, "learning_rate": 1.839210402999063e-05, - "loss": 0.1624, + "loss": 0.1851, "step": 5490 }, { "epoch": 2.5773195876288657, - "grad_norm": 2.0399646759033203, + "grad_norm": 1.7404872179031372, "learning_rate": 1.8389175257731958e-05, - "loss": 0.1572, + "loss": 0.17, "step": 5500 }, { "epoch": 2.5773195876288657, - "eval_loss": 0.03521975129842758, - "eval_pearson_cosine": 0.8123064687585853, - "eval_pearson_dot": 0.7368072149645286, - "eval_pearson_euclidean": 0.8002505346194582, - "eval_pearson_manhattan": 0.8020724209892114, - "eval_runtime": 3.2078, - "eval_samples_per_second": 467.604, - "eval_spearman_cosine": 0.8135314473340735, - "eval_spearman_dot": 0.7336319115548467, - "eval_spearman_euclidean": 0.8084064839841847, - "eval_spearman_manhattan": 0.8100456053831014, - "eval_steps_per_second": 29.303, + "eval_loss": 0.052087847143411636, + "eval_pearson_cosine": 0.7876996404495316, + "eval_pearson_dot": 0.6239791899062368, + "eval_pearson_euclidean": 0.7632876525798107, + "eval_pearson_manhattan": 0.7621358820486535, + "eval_runtime": 3.2114, + "eval_samples_per_second": 467.081, + "eval_spearman_cosine": 0.7959069730064968, + "eval_spearman_dot": 0.6246328836004849, + "eval_spearman_euclidean": 0.775279722994119, + "eval_spearman_manhattan": 0.7746119015682993, + "eval_steps_per_second": 29.27, "step": 5500 }, { "epoch": 2.5820056232427366, - "grad_norm": 1.6468899250030518, + "grad_norm": 1.7516288757324219, "learning_rate": 1.838624648547329e-05, - "loss": 0.1884, + "loss": 0.2103, "step": 5510 }, { "epoch": 2.5866916588566076, - "grad_norm": 1.9936749935150146, + "grad_norm": 2.0095157623291016, "learning_rate": 1.838331771321462e-05, - "loss": 0.1692, + "loss": 0.2149, "step": 5520 }, { "epoch": 2.591377694470478, - "grad_norm": 2.251502513885498, + "grad_norm": 2.3408539295196533, "learning_rate": 1.8380388940955953e-05, - "loss": 0.1657, + "loss": 0.1737, "step": 5530 }, { "epoch": 2.5960637300843485, - "grad_norm": 1.6334820985794067, + "grad_norm": 2.1319546699523926, "learning_rate": 1.8377460168697283e-05, - "loss": 0.1692, + "loss": 0.1963, "step": 5540 }, { "epoch": 2.600749765698219, - "grad_norm": 1.9554181098937988, + "grad_norm": 1.6769325733184814, "learning_rate": 1.8374531396438616e-05, - "loss": 0.167, + "loss": 0.1953, "step": 5550 }, { "epoch": 2.60543580131209, - "grad_norm": 1.866745114326477, + "grad_norm": 2.503340244293213, "learning_rate": 1.8371602624179945e-05, - "loss": 0.1508, + "loss": 0.1872, "step": 5560 }, { "epoch": 2.610121836925961, - "grad_norm": 1.533691644668579, + "grad_norm": 1.9402929544448853, "learning_rate": 1.8368673851921278e-05, - "loss": 0.1726, + "loss": 0.2101, "step": 5570 }, { "epoch": 2.6148078725398314, - "grad_norm": 1.8526815176010132, + "grad_norm": 1.980652928352356, "learning_rate": 1.8365745079662608e-05, - "loss": 0.1714, + "loss": 0.1847, "step": 5580 }, { "epoch": 2.619493908153702, - "grad_norm": 1.5112234354019165, + "grad_norm": 2.6117632389068604, "learning_rate": 1.8362816307403937e-05, - "loss": 0.1424, + "loss": 0.1934, "step": 5590 }, { "epoch": 2.624179943767573, - "grad_norm": 1.8528941869735718, + "grad_norm": 1.8052171468734741, "learning_rate": 1.835988753514527e-05, - "loss": 0.1677, + "loss": 0.1827, "step": 5600 }, { "epoch": 2.6288659793814433, - "grad_norm": 1.9312357902526855, + "grad_norm": 2.384441375732422, "learning_rate": 1.83569587628866e-05, - "loss": 0.1505, + "loss": 0.1765, "step": 5610 }, { "epoch": 2.633552014995314, - "grad_norm": 1.429540753364563, + "grad_norm": 1.8288124799728394, "learning_rate": 1.8354029990627932e-05, - "loss": 0.1559, + "loss": 0.189, "step": 5620 }, { "epoch": 2.6382380506091847, - "grad_norm": 1.532606840133667, + "grad_norm": 1.8904608488082886, "learning_rate": 1.8351101218369262e-05, - "loss": 0.1601, + "loss": 0.1858, "step": 5630 }, { "epoch": 2.642924086223055, - "grad_norm": 2.127380847930908, + "grad_norm": 2.4750778675079346, "learning_rate": 1.834817244611059e-05, - "loss": 0.1992, + "loss": 0.2139, "step": 5640 }, { "epoch": 2.647610121836926, - "grad_norm": 1.3522067070007324, + "grad_norm": 1.642340898513794, "learning_rate": 1.834524367385192e-05, - "loss": 0.1629, + "loss": 0.2069, "step": 5650 }, { "epoch": 2.6522961574507966, - "grad_norm": 2.4547903537750244, + "grad_norm": 2.656212091445923, "learning_rate": 1.8342314901593254e-05, - "loss": 0.1817, + "loss": 0.1975, "step": 5660 }, { "epoch": 2.6569821930646675, - "grad_norm": 1.7367674112319946, + "grad_norm": 1.8371754884719849, "learning_rate": 1.8339386129334583e-05, - "loss": 0.1764, + "loss": 0.2044, "step": 5670 }, { "epoch": 2.661668228678538, - "grad_norm": 1.256063461303711, + "grad_norm": 1.4980151653289795, "learning_rate": 1.8336457357075916e-05, - "loss": 0.1532, + "loss": 0.1641, "step": 5680 }, { "epoch": 2.6663542642924085, - "grad_norm": 1.6752853393554688, + "grad_norm": 2.300854206085205, "learning_rate": 1.8333528584817246e-05, - "loss": 0.2087, + "loss": 0.22, "step": 5690 }, { "epoch": 2.6710402999062794, - "grad_norm": 1.8749183416366577, + "grad_norm": 1.6950479745864868, "learning_rate": 1.8330599812558575e-05, - "loss": 0.1722, + "loss": 0.194, "step": 5700 }, { "epoch": 2.67572633552015, - "grad_norm": 1.9365167617797852, + "grad_norm": 2.4695844650268555, "learning_rate": 1.8327671040299908e-05, - "loss": 0.1828, + "loss": 0.2099, "step": 5710 }, { "epoch": 2.680412371134021, - "grad_norm": 1.8167498111724854, + "grad_norm": 1.6376041173934937, "learning_rate": 1.8324742268041237e-05, - "loss": 0.1566, + "loss": 0.1844, "step": 5720 }, { "epoch": 2.6850984067478914, - "grad_norm": 1.8729602098464966, + "grad_norm": 2.3736367225646973, "learning_rate": 1.832181349578257e-05, - "loss": 0.159, + "loss": 0.1886, "step": 5730 }, { "epoch": 2.689784442361762, - "grad_norm": 1.565158724784851, + "grad_norm": 1.9710888862609863, "learning_rate": 1.83188847235239e-05, - "loss": 0.1669, + "loss": 0.1881, "step": 5740 }, { "epoch": 2.6944704779756328, - "grad_norm": 2.2538883686065674, + "grad_norm": 2.38665771484375, "learning_rate": 1.8315955951265233e-05, - "loss": 0.1353, + "loss": 0.174, "step": 5750 }, { "epoch": 2.6944704779756328, - "eval_loss": 0.03330089524388313, - "eval_pearson_cosine": 0.8210346883559136, - "eval_pearson_dot": 0.7463452164747508, - "eval_pearson_euclidean": 0.8023729858931796, - "eval_pearson_manhattan": 0.8045441158759452, - "eval_runtime": 3.2103, - "eval_samples_per_second": 467.241, - "eval_spearman_cosine": 0.8210648613938134, - "eval_spearman_dot": 0.7463044894785922, - "eval_spearman_euclidean": 0.8102763562695625, - "eval_spearman_manhattan": 0.8122826147618483, - "eval_steps_per_second": 29.28, + "eval_loss": 0.052811067551374435, + "eval_pearson_cosine": 0.7875700428566148, + "eval_pearson_dot": 0.6195988091254279, + "eval_pearson_euclidean": 0.7602836109774316, + "eval_pearson_manhattan": 0.7593741208437699, + "eval_runtime": 3.3995, + "eval_samples_per_second": 441.245, + "eval_spearman_cosine": 0.7949301691809272, + "eval_spearman_dot": 0.6234473775126543, + "eval_spearman_euclidean": 0.771585595419442, + "eval_spearman_manhattan": 0.7712513119620406, + "eval_steps_per_second": 27.651, "step": 5750 }, { "epoch": 2.6991565135895033, - "grad_norm": 1.5002834796905518, + "grad_norm": 1.9556349515914917, "learning_rate": 1.8313027179006562e-05, - "loss": 0.1694, + "loss": 0.1984, "step": 5760 }, { "epoch": 2.703842549203374, - "grad_norm": 1.4530423879623413, + "grad_norm": 1.8807729482650757, "learning_rate": 1.8310098406747895e-05, - "loss": 0.1625, + "loss": 0.2057, "step": 5770 }, { "epoch": 2.7085285848172447, - "grad_norm": 1.4987956285476685, + "grad_norm": 1.74807870388031, "learning_rate": 1.8307169634489225e-05, - "loss": 0.1881, + "loss": 0.2158, "step": 5780 }, { "epoch": 2.713214620431115, - "grad_norm": 1.6835180521011353, + "grad_norm": 2.3334672451019287, "learning_rate": 1.8304240862230554e-05, - "loss": 0.1388, + "loss": 0.1676, "step": 5790 }, { "epoch": 2.717900656044986, - "grad_norm": 1.752693772315979, + "grad_norm": 2.0100314617156982, "learning_rate": 1.8301312089971887e-05, - "loss": 0.1452, + "loss": 0.1785, "step": 5800 }, { "epoch": 2.7225866916588566, - "grad_norm": 1.5680220127105713, + "grad_norm": 1.9850749969482422, "learning_rate": 1.8298383317713217e-05, - "loss": 0.1613, + "loss": 0.191, "step": 5810 }, { "epoch": 2.7272727272727275, - "grad_norm": 2.067552328109741, + "grad_norm": 2.1841342449188232, "learning_rate": 1.8295454545454546e-05, - "loss": 0.1458, + "loss": 0.1814, "step": 5820 }, { "epoch": 2.731958762886598, - "grad_norm": 2.0963711738586426, + "grad_norm": 2.6331992149353027, "learning_rate": 1.8292525773195876e-05, - "loss": 0.1972, + "loss": 0.2218, "step": 5830 }, { "epoch": 2.7366447985004685, - "grad_norm": 1.5372573137283325, + "grad_norm": 2.161170482635498, "learning_rate": 1.828959700093721e-05, - "loss": 0.1657, + "loss": 0.1834, "step": 5840 }, { "epoch": 2.7413308341143394, - "grad_norm": 2.0116796493530273, + "grad_norm": 2.3174726963043213, "learning_rate": 1.8286668228678538e-05, - "loss": 0.1576, + "loss": 0.1863, "step": 5850 }, { "epoch": 2.74601686972821, - "grad_norm": 1.3485506772994995, + "grad_norm": 1.9110337495803833, "learning_rate": 1.828373945641987e-05, - "loss": 0.1503, + "loss": 0.179, "step": 5860 }, { "epoch": 2.750702905342081, - "grad_norm": 1.7089899778366089, + "grad_norm": 2.59794545173645, "learning_rate": 1.82808106841612e-05, - "loss": 0.1787, + "loss": 0.1891, "step": 5870 }, { "epoch": 2.7553889409559513, - "grad_norm": 1.6269711256027222, + "grad_norm": 2.1763062477111816, "learning_rate": 1.8277881911902533e-05, - "loss": 0.1879, + "loss": 0.2094, "step": 5880 }, { "epoch": 2.760074976569822, - "grad_norm": 1.6839511394500732, + "grad_norm": 2.3349921703338623, "learning_rate": 1.8274953139643863e-05, - "loss": 0.1499, + "loss": 0.1848, "step": 5890 }, { "epoch": 2.7647610121836927, - "grad_norm": 1.829288363456726, + "grad_norm": 2.5409889221191406, "learning_rate": 1.8272024367385192e-05, - "loss": 0.1776, + "loss": 0.2392, "step": 5900 }, { "epoch": 2.7694470477975632, - "grad_norm": 1.4138745069503784, + "grad_norm": 2.0501179695129395, "learning_rate": 1.8269095595126525e-05, - "loss": 0.1612, + "loss": 0.2116, "step": 5910 }, { "epoch": 2.774133083411434, - "grad_norm": 1.4199497699737549, + "grad_norm": 2.486163854598999, "learning_rate": 1.8266166822867855e-05, - "loss": 0.1717, + "loss": 0.1899, "step": 5920 }, { "epoch": 2.7788191190253047, - "grad_norm": 1.8683243989944458, + "grad_norm": 2.43495774269104, "learning_rate": 1.8263238050609187e-05, - "loss": 0.1701, + "loss": 0.212, "step": 5930 }, { "epoch": 2.783505154639175, - "grad_norm": 1.5344245433807373, + "grad_norm": 1.9861983060836792, "learning_rate": 1.8260309278350517e-05, - "loss": 0.1704, + "loss": 0.2135, "step": 5940 }, { "epoch": 2.788191190253046, - "grad_norm": 1.8963440656661987, + "grad_norm": 2.46321964263916, "learning_rate": 1.825738050609185e-05, - "loss": 0.1876, + "loss": 0.2033, "step": 5950 }, { "epoch": 2.7928772258669166, - "grad_norm": 2.009709358215332, + "grad_norm": 2.2375664710998535, "learning_rate": 1.825445173383318e-05, - "loss": 0.204, + "loss": 0.2483, "step": 5960 }, { "epoch": 2.7975632614807875, - "grad_norm": 1.5217466354370117, + "grad_norm": 2.2775328159332275, "learning_rate": 1.825152296157451e-05, - "loss": 0.1582, + "loss": 0.1921, "step": 5970 }, { "epoch": 2.802249297094658, - "grad_norm": 1.7628065347671509, + "grad_norm": 2.1675665378570557, "learning_rate": 1.8248594189315842e-05, - "loss": 0.1563, + "loss": 0.1666, "step": 5980 }, { "epoch": 2.8069353327085285, - "grad_norm": 1.8212190866470337, + "grad_norm": 1.969119668006897, "learning_rate": 1.824566541705717e-05, - "loss": 0.1662, + "loss": 0.1997, "step": 5990 }, { "epoch": 2.8116213683223994, - "grad_norm": 2.4741644859313965, + "grad_norm": 3.1404953002929688, "learning_rate": 1.82427366447985e-05, - "loss": 0.1555, + "loss": 0.1896, "step": 6000 }, { "epoch": 2.8116213683223994, - "eval_loss": 0.03254643455147743, - "eval_pearson_cosine": 0.818468145503374, - "eval_pearson_dot": 0.7526356355108419, - "eval_pearson_euclidean": 0.7938927456366471, - "eval_pearson_manhattan": 0.7958636725988697, - "eval_runtime": 3.1897, - "eval_samples_per_second": 470.262, - "eval_spearman_cosine": 0.8183439432602023, - "eval_spearman_dot": 0.7537522979367676, - "eval_spearman_euclidean": 0.8019040119245933, - "eval_spearman_manhattan": 0.8036153637269691, - "eval_steps_per_second": 29.47, + "eval_loss": 0.0506160594522953, + "eval_pearson_cosine": 0.7847906228898438, + "eval_pearson_dot": 0.6051671435295134, + "eval_pearson_euclidean": 0.7605907930754796, + "eval_pearson_manhattan": 0.7595079560474858, + "eval_runtime": 3.1727, + "eval_samples_per_second": 472.783, + "eval_spearman_cosine": 0.7891227097506602, + "eval_spearman_dot": 0.6083246161085648, + "eval_spearman_euclidean": 0.7718421311199012, + "eval_spearman_manhattan": 0.7711946620901435, + "eval_steps_per_second": 29.628, "step": 6000 }, { "epoch": 2.81630740393627, - "grad_norm": 1.460700273513794, + "grad_norm": 2.5197956562042236, "learning_rate": 1.8239807872539834e-05, - "loss": 0.1745, + "loss": 0.19, "step": 6010 }, { "epoch": 2.820993439550141, - "grad_norm": 1.7924867868423462, + "grad_norm": 2.342001438140869, "learning_rate": 1.8236879100281163e-05, - "loss": 0.151, + "loss": 0.1902, "step": 6020 }, { "epoch": 2.8256794751640113, - "grad_norm": 1.7853138446807861, + "grad_norm": 2.081195592880249, "learning_rate": 1.8233950328022493e-05, - "loss": 0.1629, + "loss": 0.1861, "step": 6030 }, { "epoch": 2.830365510777882, - "grad_norm": 1.8576308488845825, + "grad_norm": 2.2222466468811035, "learning_rate": 1.8231021555763825e-05, - "loss": 0.173, + "loss": 0.1956, "step": 6040 }, { "epoch": 2.8350515463917527, - "grad_norm": 1.7521380186080933, + "grad_norm": 2.531655788421631, "learning_rate": 1.8228092783505155e-05, - "loss": 0.1741, + "loss": 0.2001, "step": 6050 }, { "epoch": 2.839737582005623, - "grad_norm": 1.6762547492980957, + "grad_norm": 1.8432163000106812, "learning_rate": 1.8225164011246488e-05, - "loss": 0.1662, + "loss": 0.1917, "step": 6060 }, { "epoch": 2.844423617619494, - "grad_norm": 1.985397219657898, + "grad_norm": 1.8071563243865967, "learning_rate": 1.8222235238987817e-05, - "loss": 0.1889, + "loss": 0.2292, "step": 6070 }, { "epoch": 2.8491096532333646, - "grad_norm": 2.0703446865081787, + "grad_norm": 2.596858501434326, "learning_rate": 1.821930646672915e-05, - "loss": 0.1772, + "loss": 0.2025, "step": 6080 }, { "epoch": 2.853795688847235, - "grad_norm": 2.128908395767212, + "grad_norm": 2.516810655593872, "learning_rate": 1.821637769447048e-05, - "loss": 0.1644, + "loss": 0.2008, "step": 6090 }, { "epoch": 2.858481724461106, - "grad_norm": 2.1612260341644287, + "grad_norm": 2.685643196105957, "learning_rate": 1.821344892221181e-05, - "loss": 0.2097, + "loss": 0.2181, "step": 6100 }, { "epoch": 2.8631677600749765, - "grad_norm": 2.220729351043701, + "grad_norm": 2.3013880252838135, "learning_rate": 1.8210520149953142e-05, - "loss": 0.1746, + "loss": 0.2089, "step": 6110 }, { "epoch": 2.8678537956888475, - "grad_norm": 2.017711639404297, + "grad_norm": 2.1551058292388916, "learning_rate": 1.820759137769447e-05, - "loss": 0.1816, + "loss": 0.2096, "step": 6120 }, { "epoch": 2.872539831302718, - "grad_norm": 1.7267364263534546, + "grad_norm": 2.2167651653289795, "learning_rate": 1.8204662605435805e-05, - "loss": 0.1738, + "loss": 0.1928, "step": 6130 }, { "epoch": 2.8772258669165884, - "grad_norm": 1.7841973304748535, + "grad_norm": 2.3424344062805176, "learning_rate": 1.8201733833177134e-05, - "loss": 0.1598, + "loss": 0.1843, "step": 6140 }, { "epoch": 2.8819119025304594, - "grad_norm": 1.9603809118270874, + "grad_norm": 2.8250808715820312, "learning_rate": 1.8198805060918464e-05, - "loss": 0.1867, + "loss": 0.2221, "step": 6150 }, { "epoch": 2.88659793814433, - "grad_norm": 2.0242254734039307, + "grad_norm": 2.3614280223846436, "learning_rate": 1.8195876288659796e-05, - "loss": 0.164, + "loss": 0.1862, "step": 6160 }, { "epoch": 2.891283973758201, - "grad_norm": 1.3930503129959106, + "grad_norm": 1.918118953704834, "learning_rate": 1.8192947516401126e-05, - "loss": 0.1469, + "loss": 0.1867, "step": 6170 }, { "epoch": 2.8959700093720713, - "grad_norm": 1.4639194011688232, + "grad_norm": 1.8377277851104736, "learning_rate": 1.8190018744142455e-05, - "loss": 0.1676, + "loss": 0.1826, "step": 6180 }, { "epoch": 2.9006560449859418, - "grad_norm": 1.9063324928283691, + "grad_norm": 2.527493715286255, "learning_rate": 1.8187089971883788e-05, - "loss": 0.1644, + "loss": 0.1986, "step": 6190 }, { "epoch": 2.9053420805998127, - "grad_norm": 1.8006539344787598, + "grad_norm": 2.125455379486084, "learning_rate": 1.8184161199625118e-05, - "loss": 0.1695, + "loss": 0.1936, "step": 6200 }, { "epoch": 2.910028116213683, - "grad_norm": 1.7670259475708008, + "grad_norm": 2.2447023391723633, "learning_rate": 1.8181232427366447e-05, - "loss": 0.1624, + "loss": 0.1887, "step": 6210 }, { "epoch": 2.914714151827554, - "grad_norm": 1.9622496366500854, + "grad_norm": 2.3427064418792725, "learning_rate": 1.817830365510778e-05, - "loss": 0.1777, + "loss": 0.197, "step": 6220 }, { "epoch": 2.9194001874414246, - "grad_norm": 1.9627212285995483, + "grad_norm": 2.480989933013916, "learning_rate": 1.817537488284911e-05, - "loss": 0.1817, + "loss": 0.2145, "step": 6230 }, { "epoch": 2.924086223055295, - "grad_norm": 1.9101510047912598, + "grad_norm": 2.2610814571380615, "learning_rate": 1.8172446110590443e-05, - "loss": 0.1743, + "loss": 0.1854, "step": 6240 }, { "epoch": 2.928772258669166, - "grad_norm": 1.890410304069519, + "grad_norm": 2.220090866088867, "learning_rate": 1.8169517338331772e-05, - "loss": 0.152, + "loss": 0.1897, "step": 6250 }, { "epoch": 2.928772258669166, - "eval_loss": 0.03263232484459877, - "eval_pearson_cosine": 0.8153769983260584, - "eval_pearson_dot": 0.7414947307896895, - "eval_pearson_euclidean": 0.7908278401422209, - "eval_pearson_manhattan": 0.7928717466268438, - "eval_runtime": 3.1717, - "eval_samples_per_second": 472.936, - "eval_spearman_cosine": 0.8150926580171975, - "eval_spearman_dot": 0.7426551131852436, - "eval_spearman_euclidean": 0.8000506096598716, - "eval_spearman_manhattan": 0.8018413708002138, - "eval_steps_per_second": 29.637, + "eval_loss": 0.05493560805916786, + "eval_pearson_cosine": 0.7819311073703084, + "eval_pearson_dot": 0.5956676516410369, + "eval_pearson_euclidean": 0.7532898404318757, + "eval_pearson_manhattan": 0.7521288710513865, + "eval_runtime": 3.273, + "eval_samples_per_second": 458.299, + "eval_spearman_cosine": 0.7902389528249426, + "eval_spearman_dot": 0.5980515739599185, + "eval_spearman_euclidean": 0.7666646754982684, + "eval_spearman_manhattan": 0.7664381416282928, + "eval_steps_per_second": 28.72, "step": 6250 }, { "epoch": 2.9334582942830365, - "grad_norm": 1.8069051504135132, + "grad_norm": 2.517014741897583, "learning_rate": 1.8166588566073105e-05, - "loss": 0.1915, + "loss": 0.2132, "step": 6260 }, { "epoch": 2.9381443298969074, - "grad_norm": 1.2744157314300537, + "grad_norm": 1.9690264463424683, "learning_rate": 1.8163659793814434e-05, - "loss": 0.1493, + "loss": 0.1756, "step": 6270 }, { "epoch": 2.942830365510778, - "grad_norm": 1.6674821376800537, + "grad_norm": 1.812945008277893, "learning_rate": 1.8160731021555767e-05, - "loss": 0.1618, + "loss": 0.1848, "step": 6280 }, { "epoch": 2.9475164011246484, - "grad_norm": 2.0318241119384766, + "grad_norm": 2.003582000732422, "learning_rate": 1.8157802249297097e-05, - "loss": 0.1769, + "loss": 0.2071, "step": 6290 }, { "epoch": 2.9522024367385193, - "grad_norm": 2.4880385398864746, + "grad_norm": 3.1531078815460205, "learning_rate": 1.8154873477038426e-05, - "loss": 0.1886, + "loss": 0.2146, "step": 6300 }, { "epoch": 2.95688847235239, - "grad_norm": 2.0214920043945312, + "grad_norm": 2.5869102478027344, "learning_rate": 1.815194470477976e-05, - "loss": 0.1806, + "loss": 0.1854, "step": 6310 }, { "epoch": 2.9615745079662608, - "grad_norm": 1.6570240259170532, + "grad_norm": 2.2422142028808594, "learning_rate": 1.814901593252109e-05, - "loss": 0.1681, + "loss": 0.179, "step": 6320 }, { "epoch": 2.9662605435801312, - "grad_norm": 1.9034634828567505, + "grad_norm": 2.346750497817993, "learning_rate": 1.814608716026242e-05, - "loss": 0.1465, + "loss": 0.1937, "step": 6330 }, { "epoch": 2.9709465791940017, - "grad_norm": 1.3237876892089844, + "grad_norm": 1.6241106986999512, "learning_rate": 1.814315838800375e-05, - "loss": 0.1658, + "loss": 0.1945, "step": 6340 }, { "epoch": 2.9756326148078727, - "grad_norm": 2.253992795944214, + "grad_norm": 2.8439619541168213, "learning_rate": 1.814022961574508e-05, - "loss": 0.1993, + "loss": 0.2265, "step": 6350 }, { "epoch": 2.980318650421743, - "grad_norm": 1.5123599767684937, + "grad_norm": 1.69356107711792, "learning_rate": 1.813730084348641e-05, - "loss": 0.1713, + "loss": 0.1869, "step": 6360 }, { "epoch": 2.985004686035614, - "grad_norm": 2.3374693393707275, + "grad_norm": 2.3386266231536865, "learning_rate": 1.8134372071227743e-05, - "loss": 0.1545, + "loss": 0.1752, "step": 6370 }, { "epoch": 2.9896907216494846, - "grad_norm": 1.5885390043258667, + "grad_norm": 1.873734712600708, "learning_rate": 1.8131443298969072e-05, - "loss": 0.1689, + "loss": 0.1972, "step": 6380 }, { "epoch": 2.994376757263355, - "grad_norm": 2.09841251373291, + "grad_norm": 2.3732969760894775, "learning_rate": 1.8128514526710402e-05, - "loss": 0.199, + "loss": 0.2229, "step": 6390 }, { "epoch": 2.999062792877226, - "grad_norm": 1.7730973958969116, + "grad_norm": 2.2805166244506836, "learning_rate": 1.8125585754451735e-05, - "loss": 0.1674, + "loss": 0.2022, "step": 6400 }, { "epoch": 3.0037488284910965, - "grad_norm": 1.6908677816390991, + "grad_norm": 1.8053455352783203, "learning_rate": 1.8122656982193064e-05, - "loss": 0.1177, + "loss": 0.132, "step": 6410 }, { "epoch": 3.0084348641049674, - "grad_norm": 1.376086711883545, + "grad_norm": 2.322193145751953, "learning_rate": 1.8119728209934397e-05, - "loss": 0.1203, + "loss": 0.1341, "step": 6420 }, { "epoch": 3.013120899718838, - "grad_norm": 1.7540253400802612, + "grad_norm": 2.2462315559387207, "learning_rate": 1.8116799437675727e-05, - "loss": 0.1133, + "loss": 0.1294, "step": 6430 }, { "epoch": 3.0178069353327084, - "grad_norm": 1.0979869365692139, + "grad_norm": 1.5314648151397705, "learning_rate": 1.811387066541706e-05, - "loss": 0.0976, + "loss": 0.0974, "step": 6440 }, { "epoch": 3.0224929709465793, - "grad_norm": 1.4160518646240234, + "grad_norm": 1.4587002992630005, "learning_rate": 1.811094189315839e-05, - "loss": 0.1084, + "loss": 0.1019, "step": 6450 }, { "epoch": 3.02717900656045, - "grad_norm": 1.0536082983016968, + "grad_norm": 0.887192964553833, "learning_rate": 1.8108013120899722e-05, - "loss": 0.0951, + "loss": 0.0945, "step": 6460 }, { "epoch": 3.0318650421743207, - "grad_norm": 1.4449706077575684, + "grad_norm": 1.8087823390960693, "learning_rate": 1.810508434864105e-05, - "loss": 0.1148, + "loss": 0.1103, "step": 6470 }, { "epoch": 3.036551077788191, - "grad_norm": 1.9929163455963135, + "grad_norm": 2.4967281818389893, "learning_rate": 1.8102155576382384e-05, - "loss": 0.1092, + "loss": 0.1261, "step": 6480 }, { "epoch": 3.0412371134020617, - "grad_norm": 1.5023281574249268, + "grad_norm": 1.4955415725708008, "learning_rate": 1.8099226804123714e-05, - "loss": 0.1366, + "loss": 0.1255, "step": 6490 }, { "epoch": 3.0459231490159326, - "grad_norm": 1.0478644371032715, + "grad_norm": 1.0758466720581055, "learning_rate": 1.8096298031865043e-05, - "loss": 0.1, + "loss": 0.105, "step": 6500 }, { "epoch": 3.0459231490159326, - "eval_loss": 0.03124774619936943, - "eval_pearson_cosine": 0.819439330096543, - "eval_pearson_dot": 0.756545112697367, - "eval_pearson_euclidean": 0.7885948455069851, - "eval_pearson_manhattan": 0.790849549554963, - "eval_runtime": 3.2845, - "eval_samples_per_second": 456.685, - "eval_spearman_cosine": 0.8189547958390648, - "eval_spearman_dot": 0.7570952217597201, - "eval_spearman_euclidean": 0.7971512626570622, - "eval_spearman_manhattan": 0.7989572580196979, - "eval_steps_per_second": 28.619, + "eval_loss": 0.04503355920314789, + "eval_pearson_cosine": 0.7887213700331159, + "eval_pearson_dot": 0.638523586851079, + "eval_pearson_euclidean": 0.7527277991826509, + "eval_pearson_manhattan": 0.7516127412435569, + "eval_runtime": 3.2262, + "eval_samples_per_second": 464.937, + "eval_spearman_cosine": 0.7931480987271526, + "eval_spearman_dot": 0.6449590450062397, + "eval_spearman_euclidean": 0.7674534939783395, + "eval_spearman_manhattan": 0.7668609796988295, + "eval_steps_per_second": 29.136, "step": 6500 }, { "epoch": 3.050609184629803, - "grad_norm": 1.7084600925445557, + "grad_norm": 2.3318862915039062, "learning_rate": 1.8093369259606376e-05, - "loss": 0.1258, + "loss": 0.1364, "step": 6510 }, { "epoch": 3.055295220243674, - "grad_norm": 1.9885269403457642, + "grad_norm": 1.8145966529846191, "learning_rate": 1.8090440487347706e-05, - "loss": 0.1275, + "loss": 0.1236, "step": 6520 }, { "epoch": 3.0599812558575445, - "grad_norm": 1.5580956935882568, + "grad_norm": 1.584845781326294, "learning_rate": 1.8087511715089035e-05, - "loss": 0.1147, + "loss": 0.128, "step": 6530 }, { "epoch": 3.064667291471415, - "grad_norm": 1.2762457132339478, + "grad_norm": 1.2164894342422485, "learning_rate": 1.8084582942830368e-05, - "loss": 0.1087, + "loss": 0.1233, "step": 6540 }, { "epoch": 3.069353327085286, - "grad_norm": 1.7397170066833496, + "grad_norm": 1.761679768562317, "learning_rate": 1.8081654170571698e-05, - "loss": 0.107, + "loss": 0.1107, "step": 6550 }, { "epoch": 3.0740393626991565, - "grad_norm": 1.6765695810317993, + "grad_norm": 1.9818497896194458, "learning_rate": 1.8078725398313027e-05, - "loss": 0.1389, + "loss": 0.1373, "step": 6560 }, { "epoch": 3.0787253983130274, - "grad_norm": 1.6627321243286133, + "grad_norm": 2.124472141265869, "learning_rate": 1.807579662605436e-05, - "loss": 0.1315, + "loss": 0.1409, "step": 6570 }, { "epoch": 3.083411433926898, - "grad_norm": 1.4617902040481567, + "grad_norm": 1.6288917064666748, "learning_rate": 1.807286785379569e-05, - "loss": 0.1003, + "loss": 0.1022, "step": 6580 }, { "epoch": 3.0880974695407684, - "grad_norm": 1.4063915014266968, + "grad_norm": 1.5154041051864624, "learning_rate": 1.806993908153702e-05, - "loss": 0.1256, + "loss": 0.1167, "step": 6590 }, { "epoch": 3.0927835051546393, - "grad_norm": 1.4291496276855469, + "grad_norm": 1.736624836921692, "learning_rate": 1.8067010309278352e-05, - "loss": 0.0952, + "loss": 0.0932, "step": 6600 }, { "epoch": 3.0974695407685098, - "grad_norm": 1.7208278179168701, + "grad_norm": 2.2292051315307617, "learning_rate": 1.806408153701968e-05, - "loss": 0.1232, + "loss": 0.1458, "step": 6610 }, { "epoch": 3.1021555763823807, - "grad_norm": 1.4286062717437744, + "grad_norm": 1.8836901187896729, "learning_rate": 1.8061152764761014e-05, - "loss": 0.1162, + "loss": 0.1145, "step": 6620 }, { "epoch": 3.106841611996251, - "grad_norm": 1.6473411321640015, + "grad_norm": 2.0602309703826904, "learning_rate": 1.8058223992502344e-05, - "loss": 0.1132, + "loss": 0.1091, "step": 6630 }, { "epoch": 3.1115276476101217, - "grad_norm": 1.6076815128326416, + "grad_norm": 1.7996163368225098, "learning_rate": 1.8055295220243677e-05, - "loss": 0.1224, + "loss": 0.1155, "step": 6640 }, { "epoch": 3.1162136832239926, - "grad_norm": 1.4402363300323486, + "grad_norm": 1.695610761642456, "learning_rate": 1.8052366447985006e-05, - "loss": 0.1149, + "loss": 0.1268, "step": 6650 }, { "epoch": 3.120899718837863, - "grad_norm": 1.2964081764221191, + "grad_norm": 1.6327197551727295, "learning_rate": 1.804943767572634e-05, - "loss": 0.1084, + "loss": 0.1157, "step": 6660 }, { "epoch": 3.125585754451734, - "grad_norm": 1.4852274656295776, + "grad_norm": 2.457167387008667, "learning_rate": 1.804650890346767e-05, - "loss": 0.1299, + "loss": 0.1348, "step": 6670 }, { "epoch": 3.1302717900656045, - "grad_norm": 1.2311766147613525, + "grad_norm": 1.1226837635040283, "learning_rate": 1.8043580131208998e-05, - "loss": 0.112, + "loss": 0.1112, "step": 6680 }, { "epoch": 3.134957825679475, - "grad_norm": 1.7571637630462646, + "grad_norm": 1.904137134552002, "learning_rate": 1.804065135895033e-05, - "loss": 0.1128, + "loss": 0.1288, "step": 6690 }, { "epoch": 3.139643861293346, - "grad_norm": 2.1391453742980957, + "grad_norm": 1.8711892366409302, "learning_rate": 1.803772258669166e-05, - "loss": 0.1323, + "loss": 0.1341, "step": 6700 }, { "epoch": 3.1443298969072164, - "grad_norm": 1.5623685121536255, + "grad_norm": 1.8008291721343994, "learning_rate": 1.803479381443299e-05, - "loss": 0.1139, + "loss": 0.1198, "step": 6710 }, { "epoch": 3.1490159325210874, - "grad_norm": 1.7942404747009277, + "grad_norm": 1.8362482786178589, "learning_rate": 1.8031865042174323e-05, - "loss": 0.1426, + "loss": 0.1479, "step": 6720 }, { "epoch": 3.153701968134958, - "grad_norm": 1.410545825958252, + "grad_norm": 1.6150809526443481, "learning_rate": 1.8028936269915652e-05, - "loss": 0.1056, + "loss": 0.1138, "step": 6730 }, { "epoch": 3.1583880037488283, - "grad_norm": 1.8002315759658813, + "grad_norm": 2.2239766120910645, "learning_rate": 1.8026007497656982e-05, - "loss": 0.1256, + "loss": 0.1309, "step": 6740 }, { "epoch": 3.1630740393626993, - "grad_norm": 1.1960804462432861, + "grad_norm": 1.669233798980713, "learning_rate": 1.8023078725398315e-05, - "loss": 0.1075, + "loss": 0.1055, "step": 6750 }, { "epoch": 3.1630740393626993, - "eval_loss": 0.031751763075590134, - "eval_pearson_cosine": 0.818392251131673, - "eval_pearson_dot": 0.7566814090378315, - "eval_pearson_euclidean": 0.7928479508679516, - "eval_pearson_manhattan": 0.7948564653512236, - "eval_runtime": 3.2248, - "eval_samples_per_second": 465.151, - "eval_spearman_cosine": 0.8181084179548533, - "eval_spearman_dot": 0.7582501700895796, - "eval_spearman_euclidean": 0.801587599060446, - "eval_spearman_manhattan": 0.8030998382251602, - "eval_steps_per_second": 29.149, + "eval_loss": 0.04599650204181671, + "eval_pearson_cosine": 0.7875010466521104, + "eval_pearson_dot": 0.6256249285068609, + "eval_pearson_euclidean": 0.752459338375715, + "eval_pearson_manhattan": 0.7515050600000066, + "eval_runtime": 3.3759, + "eval_samples_per_second": 444.332, + "eval_spearman_cosine": 0.792665874705252, + "eval_spearman_dot": 0.6331527343675145, + "eval_spearman_euclidean": 0.7657151594858318, + "eval_spearman_manhattan": 0.7651890994431135, + "eval_steps_per_second": 27.845, "step": 6750 }, { "epoch": 3.1677600749765698, - "grad_norm": 1.8303911685943604, + "grad_norm": 2.1133861541748047, "learning_rate": 1.8020149953139644e-05, - "loss": 0.1148, + "loss": 0.1128, "step": 6760 }, { "epoch": 3.1724461105904407, - "grad_norm": 0.8495842814445496, + "grad_norm": 1.5470513105392456, "learning_rate": 1.8017221180880977e-05, - "loss": 0.1118, + "loss": 0.1246, "step": 6770 }, { "epoch": 3.177132146204311, - "grad_norm": 1.8021464347839355, + "grad_norm": 1.9087189435958862, "learning_rate": 1.8014292408622307e-05, - "loss": 0.1171, + "loss": 0.1137, "step": 6780 }, { "epoch": 3.1818181818181817, - "grad_norm": 1.4046647548675537, + "grad_norm": 1.9185723066329956, "learning_rate": 1.8011363636363636e-05, - "loss": 0.1126, + "loss": 0.1196, "step": 6790 }, { "epoch": 3.1865042174320526, - "grad_norm": 1.425398349761963, + "grad_norm": 1.7757681608200073, "learning_rate": 1.800843486410497e-05, - "loss": 0.0897, + "loss": 0.0978, "step": 6800 }, { "epoch": 3.191190253045923, - "grad_norm": 1.938445806503296, + "grad_norm": 1.6667931079864502, "learning_rate": 1.80055060918463e-05, - "loss": 0.1173, + "loss": 0.113, "step": 6810 }, { "epoch": 3.195876288659794, - "grad_norm": 1.9677798748016357, + "grad_norm": 2.101773500442505, "learning_rate": 1.800257731958763e-05, - "loss": 0.1113, + "loss": 0.1034, "step": 6820 }, { "epoch": 3.2005623242736645, - "grad_norm": 1.0695815086364746, + "grad_norm": 1.6078181266784668, "learning_rate": 1.799964854732896e-05, - "loss": 0.1251, + "loss": 0.1276, "step": 6830 }, { "epoch": 3.205248359887535, - "grad_norm": 1.5036897659301758, + "grad_norm": 1.3552807569503784, "learning_rate": 1.7996719775070294e-05, - "loss": 0.1077, + "loss": 0.1184, "step": 6840 }, { "epoch": 3.209934395501406, - "grad_norm": 1.0521790981292725, + "grad_norm": 1.5063031911849976, "learning_rate": 1.7993791002811623e-05, - "loss": 0.1121, + "loss": 0.1367, "step": 6850 }, { "epoch": 3.2146204311152764, - "grad_norm": 1.6036179065704346, + "grad_norm": 1.9200366735458374, "learning_rate": 1.7990862230552956e-05, - "loss": 0.1087, + "loss": 0.1222, "step": 6860 }, { "epoch": 3.2193064667291473, - "grad_norm": 1.5137251615524292, + "grad_norm": 1.4902859926223755, "learning_rate": 1.7987933458294286e-05, - "loss": 0.1227, + "loss": 0.1414, "step": 6870 }, { "epoch": 3.223992502343018, - "grad_norm": 1.7321174144744873, + "grad_norm": 1.7316768169403076, "learning_rate": 1.7985004686035615e-05, - "loss": 0.1165, + "loss": 0.1451, "step": 6880 }, { "epoch": 3.2286785379568883, - "grad_norm": 1.744694709777832, + "grad_norm": 1.8446965217590332, "learning_rate": 1.7982075913776945e-05, - "loss": 0.1164, + "loss": 0.1227, "step": 6890 }, { "epoch": 3.2333645735707592, - "grad_norm": 1.5463277101516724, + "grad_norm": 1.5571191310882568, "learning_rate": 1.7979147141518278e-05, - "loss": 0.1047, + "loss": 0.1193, "step": 6900 }, { "epoch": 3.2380506091846297, - "grad_norm": 1.663386583328247, + "grad_norm": 2.3433127403259277, "learning_rate": 1.7976218369259607e-05, - "loss": 0.102, + "loss": 0.1175, "step": 6910 }, { "epoch": 3.2427366447985007, - "grad_norm": 1.3090193271636963, + "grad_norm": 1.8916155099868774, "learning_rate": 1.7973289597000937e-05, - "loss": 0.1206, + "loss": 0.1352, "step": 6920 }, { "epoch": 3.247422680412371, - "grad_norm": 1.535120964050293, + "grad_norm": 1.3708059787750244, "learning_rate": 1.797036082474227e-05, - "loss": 0.1131, + "loss": 0.1174, "step": 6930 }, { "epoch": 3.2521087160262416, - "grad_norm": 2.086014986038208, + "grad_norm": 1.5119550228118896, "learning_rate": 1.79674320524836e-05, - "loss": 0.1237, + "loss": 0.1331, "step": 6940 }, { "epoch": 3.2567947516401126, - "grad_norm": 1.6445001363754272, + "grad_norm": 1.8884799480438232, "learning_rate": 1.7964503280224932e-05, - "loss": 0.1079, + "loss": 0.125, "step": 6950 }, { "epoch": 3.261480787253983, - "grad_norm": 1.4344754219055176, + "grad_norm": 1.6222647428512573, "learning_rate": 1.796157450796626e-05, - "loss": 0.1077, + "loss": 0.1191, "step": 6960 }, { "epoch": 3.266166822867854, - "grad_norm": 0.9027751088142395, + "grad_norm": 1.2746548652648926, "learning_rate": 1.7958645735707594e-05, - "loss": 0.0977, + "loss": 0.1256, "step": 6970 }, { "epoch": 3.2708528584817245, - "grad_norm": 1.6101003885269165, + "grad_norm": 1.883239984512329, "learning_rate": 1.7955716963448924e-05, - "loss": 0.1233, + "loss": 0.1283, "step": 6980 }, { "epoch": 3.275538894095595, - "grad_norm": 1.6779502630233765, + "grad_norm": 2.0994622707366943, "learning_rate": 1.7952788191190253e-05, - "loss": 0.1259, + "loss": 0.1316, "step": 6990 }, { "epoch": 3.280224929709466, - "grad_norm": 1.1465091705322266, + "grad_norm": 1.844297170639038, "learning_rate": 1.7949859418931586e-05, - "loss": 0.0971, + "loss": 0.1145, "step": 7000 }, { "epoch": 3.280224929709466, - "eval_loss": 0.031160470098257065, - "eval_pearson_cosine": 0.8183441854138493, - "eval_pearson_dot": 0.756100708061723, - "eval_pearson_euclidean": 0.7881792865072583, - "eval_pearson_manhattan": 0.7905067817277995, - "eval_runtime": 3.3694, - "eval_samples_per_second": 445.183, - "eval_spearman_cosine": 0.8175806864862152, - "eval_spearman_dot": 0.7571609305062721, - "eval_spearman_euclidean": 0.7969525723417172, - "eval_spearman_manhattan": 0.7992497485259732, - "eval_steps_per_second": 27.898, + "eval_loss": 0.045286137610673904, + "eval_pearson_cosine": 0.7924534055849843, + "eval_pearson_dot": 0.6315508963322145, + "eval_pearson_euclidean": 0.7559423098506386, + "eval_pearson_manhattan": 0.7547566595358202, + "eval_runtime": 3.1537, + "eval_samples_per_second": 475.63, + "eval_spearman_cosine": 0.7977173793324885, + "eval_spearman_dot": 0.6408281685024612, + "eval_spearman_euclidean": 0.7678066060450297, + "eval_spearman_manhattan": 0.7671464837523794, + "eval_steps_per_second": 29.806, "step": 7000 - }, - { - "epoch": 3.2849109653233364, - "grad_norm": 1.3195544481277466, - "learning_rate": 1.7946930646672916e-05, - "loss": 0.1136, - "step": 7010 - }, - { - "epoch": 3.2895970009372073, - "grad_norm": 1.7174955606460571, - "learning_rate": 1.794400187441425e-05, - "loss": 0.1114, - "step": 7020 - }, - { - "epoch": 3.294283036551078, - "grad_norm": 1.8189715147018433, - "learning_rate": 1.7941073102155578e-05, - "loss": 0.133, - "step": 7030 - }, - { - "epoch": 3.2989690721649483, - "grad_norm": 1.8281652927398682, - "learning_rate": 1.793814432989691e-05, - "loss": 0.1079, - "step": 7040 - }, - { - "epoch": 3.303655107778819, - "grad_norm": 1.686579704284668, - "learning_rate": 1.793521555763824e-05, - "loss": 0.1038, - "step": 7050 - }, - { - "epoch": 3.3083411433926897, - "grad_norm": 1.5126315355300903, - "learning_rate": 1.793228678537957e-05, - "loss": 0.1181, - "step": 7060 - }, - { - "epoch": 3.3130271790065606, - "grad_norm": 1.5008283853530884, - "learning_rate": 1.79293580131209e-05, - "loss": 0.1223, - "step": 7070 - }, - { - "epoch": 3.317713214620431, - "grad_norm": 0.7563474178314209, - "learning_rate": 1.7926429240862232e-05, - "loss": 0.1028, - "step": 7080 - }, - { - "epoch": 3.3223992502343016, - "grad_norm": 0.9533982872962952, - "learning_rate": 1.7923500468603562e-05, - "loss": 0.1032, - "step": 7090 - }, - { - "epoch": 3.3270852858481725, - "grad_norm": 1.5246295928955078, - "learning_rate": 1.792057169634489e-05, - "loss": 0.128, - "step": 7100 - }, - { - "epoch": 3.331771321462043, - "grad_norm": 1.4202959537506104, - "learning_rate": 1.7917642924086224e-05, - "loss": 0.1173, - "step": 7110 - }, - { - "epoch": 3.336457357075914, - "grad_norm": 1.5270695686340332, - "learning_rate": 1.7914714151827554e-05, - "loss": 0.1188, - "step": 7120 - }, - { - "epoch": 3.3411433926897844, - "grad_norm": 1.5747121572494507, - "learning_rate": 1.7911785379568887e-05, - "loss": 0.1154, - "step": 7130 - }, - { - "epoch": 3.345829428303655, - "grad_norm": 1.4199724197387695, - "learning_rate": 1.7908856607310216e-05, - "loss": 0.1174, - "step": 7140 - }, - { - "epoch": 3.350515463917526, - "grad_norm": 1.601879596710205, - "learning_rate": 1.790592783505155e-05, - "loss": 0.1095, - "step": 7150 - }, - { - "epoch": 3.3552014995313963, - "grad_norm": 1.8753916025161743, - "learning_rate": 1.790299906279288e-05, - "loss": 0.1137, - "step": 7160 - }, - { - "epoch": 3.3598875351452673, - "grad_norm": 2.2938902378082275, - "learning_rate": 1.790007029053421e-05, - "loss": 0.1187, - "step": 7170 - }, - { - "epoch": 3.3645735707591378, - "grad_norm": 1.0570533275604248, - "learning_rate": 1.789714151827554e-05, - "loss": 0.1122, - "step": 7180 - }, - { - "epoch": 3.3692596063730083, - "grad_norm": 1.3693780899047852, - "learning_rate": 1.789421274601687e-05, - "loss": 0.1056, - "step": 7190 - }, - { - "epoch": 3.373945641986879, - "grad_norm": 1.088611364364624, - "learning_rate": 1.7891283973758203e-05, - "loss": 0.0943, - "step": 7200 - }, - { - "epoch": 3.3786316776007497, - "grad_norm": 1.651551604270935, - "learning_rate": 1.7888355201499533e-05, - "loss": 0.1148, - "step": 7210 - }, - { - "epoch": 3.3833177132146206, - "grad_norm": 1.9416500329971313, - "learning_rate": 1.7885426429240866e-05, - "loss": 0.1162, - "step": 7220 - }, - { - "epoch": 3.388003748828491, - "grad_norm": 1.549742341041565, - "learning_rate": 1.7882497656982195e-05, - "loss": 0.0984, - "step": 7230 - }, - { - "epoch": 3.3926897844423616, - "grad_norm": 1.9772891998291016, - "learning_rate": 1.7879568884723525e-05, - "loss": 0.1146, - "step": 7240 - }, - { - "epoch": 3.3973758200562325, - "grad_norm": 1.915805459022522, - "learning_rate": 1.7876640112464858e-05, - "loss": 0.12, - "step": 7250 - }, - { - "epoch": 3.3973758200562325, - "eval_loss": 0.030257537961006165, - "eval_pearson_cosine": 0.8236641701952188, - "eval_pearson_dot": 0.7683165917501924, - "eval_pearson_euclidean": 0.7929510591670237, - "eval_pearson_manhattan": 0.7953417991908651, - "eval_runtime": 3.1268, - "eval_samples_per_second": 479.717, - "eval_spearman_cosine": 0.8229564695132245, - "eval_spearman_dot": 0.7689794551234463, - "eval_spearman_euclidean": 0.8016168710764218, - "eval_spearman_manhattan": 0.8035095618864339, - "eval_steps_per_second": 30.062, - "step": 7250 - }, - { - "epoch": 3.402061855670103, - "grad_norm": 1.2766612768173218, - "learning_rate": 1.7873711340206187e-05, - "loss": 0.1256, - "step": 7260 - }, - { - "epoch": 3.406747891283974, - "grad_norm": 1.6186020374298096, - "learning_rate": 1.7870782567947517e-05, - "loss": 0.106, - "step": 7270 - }, - { - "epoch": 3.4114339268978444, - "grad_norm": 1.5091776847839355, - "learning_rate": 1.7867853795688846e-05, - "loss": 0.1157, - "step": 7280 - }, - { - "epoch": 3.416119962511715, - "grad_norm": 1.5809932947158813, - "learning_rate": 1.786492502343018e-05, - "loss": 0.1151, - "step": 7290 - }, - { - "epoch": 3.420805998125586, - "grad_norm": 1.6943892240524292, - "learning_rate": 1.786199625117151e-05, - "loss": 0.1131, - "step": 7300 - }, - { - "epoch": 3.4254920337394563, - "grad_norm": 1.5174516439437866, - "learning_rate": 1.785906747891284e-05, - "loss": 0.0843, - "step": 7310 - }, - { - "epoch": 3.4301780693533273, - "grad_norm": 1.4588013887405396, - "learning_rate": 1.785613870665417e-05, - "loss": 0.1079, - "step": 7320 - }, - { - "epoch": 3.4348641049671977, - "grad_norm": 1.5333396196365356, - "learning_rate": 1.7853209934395504e-05, - "loss": 0.1309, - "step": 7330 - }, - { - "epoch": 3.4395501405810682, - "grad_norm": 1.5144264698028564, - "learning_rate": 1.7850281162136833e-05, - "loss": 0.1006, - "step": 7340 - }, - { - "epoch": 3.444236176194939, - "grad_norm": 1.2702832221984863, - "learning_rate": 1.7847352389878166e-05, - "loss": 0.1173, - "step": 7350 - }, - { - "epoch": 3.4489222118088096, - "grad_norm": 1.808031678199768, - "learning_rate": 1.7844423617619496e-05, - "loss": 0.1321, - "step": 7360 - }, - { - "epoch": 3.4536082474226806, - "grad_norm": 1.8384732007980347, - "learning_rate": 1.784149484536083e-05, - "loss": 0.1093, - "step": 7370 - }, - { - "epoch": 3.458294283036551, - "grad_norm": 1.468562126159668, - "learning_rate": 1.7838566073102158e-05, - "loss": 0.1117, - "step": 7380 - }, - { - "epoch": 3.4629803186504216, - "grad_norm": 1.647537350654602, - "learning_rate": 1.7835637300843487e-05, - "loss": 0.1161, - "step": 7390 - }, - { - "epoch": 3.4676663542642925, - "grad_norm": 1.5217511653900146, - "learning_rate": 1.783270852858482e-05, - "loss": 0.1287, - "step": 7400 - }, - { - "epoch": 3.472352389878163, - "grad_norm": 1.3853079080581665, - "learning_rate": 1.782977975632615e-05, - "loss": 0.1273, - "step": 7410 - }, - { - "epoch": 3.477038425492034, - "grad_norm": 1.938225269317627, - "learning_rate": 1.782685098406748e-05, - "loss": 0.1381, - "step": 7420 - }, - { - "epoch": 3.4817244611059044, - "grad_norm": 1.4750497341156006, - "learning_rate": 1.7823922211808812e-05, - "loss": 0.1173, - "step": 7430 - }, - { - "epoch": 3.486410496719775, - "grad_norm": 1.421303391456604, - "learning_rate": 1.7820993439550142e-05, - "loss": 0.104, - "step": 7440 - }, - { - "epoch": 3.491096532333646, - "grad_norm": 1.1640669107437134, - "learning_rate": 1.781806466729147e-05, - "loss": 0.1018, - "step": 7450 - }, - { - "epoch": 3.4957825679475163, - "grad_norm": 1.3329296112060547, - "learning_rate": 1.7815135895032804e-05, - "loss": 0.1127, - "step": 7460 - }, - { - "epoch": 3.5004686035613872, - "grad_norm": 2.0254921913146973, - "learning_rate": 1.7812207122774134e-05, - "loss": 0.1361, - "step": 7470 - }, - { - "epoch": 3.5051546391752577, - "grad_norm": 1.3234291076660156, - "learning_rate": 1.7809278350515463e-05, - "loss": 0.1211, - "step": 7480 - }, - { - "epoch": 3.509840674789128, - "grad_norm": 1.448867678642273, - "learning_rate": 1.7806349578256796e-05, - "loss": 0.0999, - "step": 7490 - }, - { - "epoch": 3.514526710402999, - "grad_norm": 1.56028151512146, - "learning_rate": 1.7803420805998125e-05, - "loss": 0.1003, - "step": 7500 - }, - { - "epoch": 3.514526710402999, - "eval_loss": 0.031522952020168304, - "eval_pearson_cosine": 0.81808882827319, - "eval_pearson_dot": 0.750183242563736, - "eval_pearson_euclidean": 0.7941389529600116, - "eval_pearson_manhattan": 0.7963593749181541, - "eval_runtime": 3.4063, - "eval_samples_per_second": 440.357, - "eval_spearman_cosine": 0.8172168876953952, - "eval_spearman_dot": 0.7505334150938179, - "eval_spearman_euclidean": 0.8027508928771647, - "eval_spearman_manhattan": 0.8047390955898335, - "eval_steps_per_second": 27.596, - "step": 7500 - }, - { - "epoch": 3.5192127460168696, - "grad_norm": 1.775154948234558, - "learning_rate": 1.780049203373946e-05, - "loss": 0.1058, - "step": 7510 - }, - { - "epoch": 3.5238987816307406, - "grad_norm": 2.034623622894287, - "learning_rate": 1.7797563261480788e-05, - "loss": 0.1198, - "step": 7520 - }, - { - "epoch": 3.528584817244611, - "grad_norm": 2.037757635116577, - "learning_rate": 1.779463448922212e-05, - "loss": 0.1047, - "step": 7530 - }, - { - "epoch": 3.5332708528584815, - "grad_norm": 1.7488856315612793, - "learning_rate": 1.779170571696345e-05, - "loss": 0.1096, - "step": 7540 - }, - { - "epoch": 3.5379568884723525, - "grad_norm": 1.4379994869232178, - "learning_rate": 1.7788776944704783e-05, - "loss": 0.1294, - "step": 7550 - }, - { - "epoch": 3.542642924086223, - "grad_norm": 1.6399285793304443, - "learning_rate": 1.7785848172446113e-05, - "loss": 0.1022, - "step": 7560 - }, - { - "epoch": 3.547328959700094, - "grad_norm": 2.018709182739258, - "learning_rate": 1.7782919400187446e-05, - "loss": 0.1109, - "step": 7570 - }, - { - "epoch": 3.5520149953139644, - "grad_norm": 0.8838659524917603, - "learning_rate": 1.7779990627928775e-05, - "loss": 0.1174, - "step": 7580 - }, - { - "epoch": 3.556701030927835, - "grad_norm": 2.33553147315979, - "learning_rate": 1.7777061855670105e-05, - "loss": 0.1303, - "step": 7590 - }, - { - "epoch": 3.561387066541706, - "grad_norm": 1.3626232147216797, - "learning_rate": 1.7774133083411434e-05, - "loss": 0.1261, - "step": 7600 - }, - { - "epoch": 3.5660731021555763, - "grad_norm": 1.8817883729934692, - "learning_rate": 1.7771204311152767e-05, - "loss": 0.129, - "step": 7610 - }, - { - "epoch": 3.570759137769447, - "grad_norm": 1.817522644996643, - "learning_rate": 1.7768275538894096e-05, - "loss": 0.1029, - "step": 7620 - }, - { - "epoch": 3.5754451733833177, - "grad_norm": 2.126809597015381, - "learning_rate": 1.7765346766635426e-05, - "loss": 0.133, - "step": 7630 - }, - { - "epoch": 3.580131208997188, - "grad_norm": 1.5726145505905151, - "learning_rate": 1.776241799437676e-05, - "loss": 0.1174, - "step": 7640 - }, - { - "epoch": 3.584817244611059, - "grad_norm": 1.2460345029830933, - "learning_rate": 1.7759489222118088e-05, - "loss": 0.104, - "step": 7650 - }, - { - "epoch": 3.5895032802249296, - "grad_norm": 0.8555458784103394, - "learning_rate": 1.775656044985942e-05, - "loss": 0.0922, - "step": 7660 - }, - { - "epoch": 3.5941893158388005, - "grad_norm": 1.0801823139190674, - "learning_rate": 1.775363167760075e-05, - "loss": 0.1155, - "step": 7670 - }, - { - "epoch": 3.598875351452671, - "grad_norm": 1.5534123182296753, - "learning_rate": 1.775070290534208e-05, - "loss": 0.1188, - "step": 7680 - }, - { - "epoch": 3.6035613870665415, - "grad_norm": 1.877568244934082, - "learning_rate": 1.7747774133083413e-05, - "loss": 0.1247, - "step": 7690 - }, - { - "epoch": 3.6082474226804124, - "grad_norm": 1.3344906568527222, - "learning_rate": 1.7744845360824743e-05, - "loss": 0.1026, - "step": 7700 - }, - { - "epoch": 3.612933458294283, - "grad_norm": 1.3001571893692017, - "learning_rate": 1.7741916588566075e-05, - "loss": 0.1084, - "step": 7710 - }, - { - "epoch": 3.617619493908154, - "grad_norm": 1.6098321676254272, - "learning_rate": 1.7738987816307405e-05, - "loss": 0.1198, - "step": 7720 - }, - { - "epoch": 3.6223055295220243, - "grad_norm": 1.9912426471710205, - "learning_rate": 1.7736059044048738e-05, - "loss": 0.1207, - "step": 7730 - }, - { - "epoch": 3.626991565135895, - "grad_norm": 1.3719514608383179, - "learning_rate": 1.7733130271790067e-05, - "loss": 0.1146, - "step": 7740 - }, - { - "epoch": 3.6316776007497658, - "grad_norm": 1.8076539039611816, - "learning_rate": 1.77302014995314e-05, - "loss": 0.1237, - "step": 7750 - }, - { - "epoch": 3.6316776007497658, - "eval_loss": 0.030823856592178345, - "eval_pearson_cosine": 0.8190383769297682, - "eval_pearson_dot": 0.7588954827942871, - "eval_pearson_euclidean": 0.7885622153014467, - "eval_pearson_manhattan": 0.7914609723187311, - "eval_runtime": 3.3684, - "eval_samples_per_second": 445.315, - "eval_spearman_cosine": 0.8178038191764202, - "eval_spearman_dot": 0.7582507405982527, - "eval_spearman_euclidean": 0.7968649602561575, - "eval_spearman_manhattan": 0.7990160021869986, - "eval_steps_per_second": 27.906, - "step": 7750 - }, - { - "epoch": 3.6363636363636362, - "grad_norm": 1.93159818649292, - "learning_rate": 1.772727272727273e-05, - "loss": 0.1406, - "step": 7760 - }, - { - "epoch": 3.641049671977507, - "grad_norm": 1.5332342386245728, - "learning_rate": 1.772434395501406e-05, - "loss": 0.1012, - "step": 7770 - }, - { - "epoch": 3.6457357075913777, - "grad_norm": 1.2360097169876099, - "learning_rate": 1.772141518275539e-05, - "loss": 0.1226, - "step": 7780 - }, - { - "epoch": 3.650421743205248, - "grad_norm": 1.1280817985534668, - "learning_rate": 1.771848641049672e-05, - "loss": 0.1096, - "step": 7790 - }, - { - "epoch": 3.655107778819119, - "grad_norm": 1.785524606704712, - "learning_rate": 1.771555763823805e-05, - "loss": 0.0999, - "step": 7800 - }, - { - "epoch": 3.6597938144329896, - "grad_norm": 1.4114161729812622, - "learning_rate": 1.771262886597938e-05, - "loss": 0.1516, - "step": 7810 - }, - { - "epoch": 3.6644798500468605, - "grad_norm": 0.764554500579834, - "learning_rate": 1.7709700093720713e-05, - "loss": 0.1187, - "step": 7820 - }, - { - "epoch": 3.669165885660731, - "grad_norm": 1.6357208490371704, - "learning_rate": 1.7706771321462043e-05, - "loss": 0.125, - "step": 7830 - }, - { - "epoch": 3.6738519212746015, - "grad_norm": 1.6706812381744385, - "learning_rate": 1.7703842549203376e-05, - "loss": 0.1231, - "step": 7840 - }, - { - "epoch": 3.6785379568884724, - "grad_norm": 1.300350308418274, - "learning_rate": 1.7700913776944705e-05, - "loss": 0.1322, - "step": 7850 - }, - { - "epoch": 3.683223992502343, - "grad_norm": 2.3312079906463623, - "learning_rate": 1.7697985004686038e-05, - "loss": 0.1215, - "step": 7860 - }, - { - "epoch": 3.687910028116214, - "grad_norm": 1.8526086807250977, - "learning_rate": 1.7695056232427368e-05, - "loss": 0.1221, - "step": 7870 - }, - { - "epoch": 3.6925960637300843, - "grad_norm": 1.1056420803070068, - "learning_rate": 1.7692127460168697e-05, - "loss": 0.1133, - "step": 7880 - }, - { - "epoch": 3.697282099343955, - "grad_norm": 1.6572386026382446, - "learning_rate": 1.768919868791003e-05, - "loss": 0.1088, - "step": 7890 - }, - { - "epoch": 3.7019681349578257, - "grad_norm": 2.0055267810821533, - "learning_rate": 1.768626991565136e-05, - "loss": 0.1329, - "step": 7900 - }, - { - "epoch": 3.706654170571696, - "grad_norm": 1.336578130722046, - "learning_rate": 1.7683341143392693e-05, - "loss": 0.1383, - "step": 7910 - }, - { - "epoch": 3.711340206185567, - "grad_norm": 1.4480973482131958, - "learning_rate": 1.7680412371134022e-05, - "loss": 0.1315, - "step": 7920 - }, - { - "epoch": 3.7160262417994376, - "grad_norm": 1.1693243980407715, - "learning_rate": 1.7677483598875355e-05, - "loss": 0.1103, - "step": 7930 - }, - { - "epoch": 3.720712277413308, - "grad_norm": 1.479332447052002, - "learning_rate": 1.7674554826616684e-05, - "loss": 0.1113, - "step": 7940 - }, - { - "epoch": 3.725398313027179, - "grad_norm": 1.9750406742095947, - "learning_rate": 1.7671626054358014e-05, - "loss": 0.1241, - "step": 7950 - }, - { - "epoch": 3.7300843486410495, - "grad_norm": 1.7816461324691772, - "learning_rate": 1.7668697282099347e-05, - "loss": 0.1437, - "step": 7960 - }, - { - "epoch": 3.7347703842549205, - "grad_norm": 1.6916552782058716, - "learning_rate": 1.7665768509840676e-05, - "loss": 0.1114, - "step": 7970 - }, - { - "epoch": 3.739456419868791, - "grad_norm": 2.0555579662323, - "learning_rate": 1.7662839737582006e-05, - "loss": 0.1294, - "step": 7980 - }, - { - "epoch": 3.7441424554826614, - "grad_norm": 1.6788338422775269, - "learning_rate": 1.7659910965323335e-05, - "loss": 0.1097, - "step": 7990 - }, - { - "epoch": 3.7488284910965324, - "grad_norm": 1.449519395828247, - "learning_rate": 1.7656982193064668e-05, - "loss": 0.0991, - "step": 8000 - }, - { - "epoch": 3.7488284910965324, - "eval_loss": 0.031480398029088974, - "eval_pearson_cosine": 0.8186177727574986, - "eval_pearson_dot": 0.7539932234321896, - "eval_pearson_euclidean": 0.7925447531882952, - "eval_pearson_manhattan": 0.7952092015244716, - "eval_runtime": 3.1967, - "eval_samples_per_second": 469.239, - "eval_spearman_cosine": 0.8171537163292079, - "eval_spearman_dot": 0.7531451005912094, - "eval_spearman_euclidean": 0.8000355030273796, - "eval_spearman_manhattan": 0.8023527525471515, - "eval_steps_per_second": 29.406, - "step": 8000 - }, - { - "epoch": 3.753514526710403, - "grad_norm": 1.525901198387146, - "learning_rate": 1.7654053420805998e-05, - "loss": 0.1211, - "step": 8010 - }, - { - "epoch": 3.758200562324274, - "grad_norm": 2.8532297611236572, - "learning_rate": 1.765112464854733e-05, - "loss": 0.1165, - "step": 8020 - }, - { - "epoch": 3.7628865979381443, - "grad_norm": 1.453282117843628, - "learning_rate": 1.764819587628866e-05, - "loss": 0.1293, - "step": 8030 - }, - { - "epoch": 3.7675726335520148, - "grad_norm": 1.6476629972457886, - "learning_rate": 1.7645267104029993e-05, - "loss": 0.1104, - "step": 8040 - }, - { - "epoch": 3.7722586691658857, - "grad_norm": 1.583380103111267, - "learning_rate": 1.7642338331771322e-05, - "loss": 0.1191, - "step": 8050 - }, - { - "epoch": 3.776944704779756, - "grad_norm": 1.4234002828598022, - "learning_rate": 1.7639409559512655e-05, - "loss": 0.1201, - "step": 8060 - }, - { - "epoch": 3.781630740393627, - "grad_norm": 2.0201187133789062, - "learning_rate": 1.7636480787253985e-05, - "loss": 0.1398, - "step": 8070 - }, - { - "epoch": 3.7863167760074976, - "grad_norm": 1.8647639751434326, - "learning_rate": 1.7633552014995314e-05, - "loss": 0.1344, - "step": 8080 - }, - { - "epoch": 3.791002811621368, - "grad_norm": 1.5310838222503662, - "learning_rate": 1.7630623242736647e-05, - "loss": 0.118, - "step": 8090 - }, - { - "epoch": 3.795688847235239, - "grad_norm": 1.740401268005371, - "learning_rate": 1.7627694470477977e-05, - "loss": 0.1187, - "step": 8100 - }, - { - "epoch": 3.8003748828491095, - "grad_norm": 1.0978221893310547, - "learning_rate": 1.762476569821931e-05, - "loss": 0.1218, - "step": 8110 - }, - { - "epoch": 3.8050609184629804, - "grad_norm": 1.8838212490081787, - "learning_rate": 1.762183692596064e-05, - "loss": 0.1349, - "step": 8120 - }, - { - "epoch": 3.809746954076851, - "grad_norm": 1.8535819053649902, - "learning_rate": 1.761890815370197e-05, - "loss": 0.118, - "step": 8130 - }, - { - "epoch": 3.8144329896907214, - "grad_norm": 1.482851266860962, - "learning_rate": 1.76159793814433e-05, - "loss": 0.1136, - "step": 8140 - }, - { - "epoch": 3.8191190253045924, - "grad_norm": 1.4817432165145874, - "learning_rate": 1.761305060918463e-05, - "loss": 0.1144, - "step": 8150 - }, - { - "epoch": 3.823805060918463, - "grad_norm": 1.8795218467712402, - "learning_rate": 1.761012183692596e-05, - "loss": 0.1337, - "step": 8160 - }, - { - "epoch": 3.8284910965323338, - "grad_norm": 1.5762320756912231, - "learning_rate": 1.7607193064667293e-05, - "loss": 0.1186, - "step": 8170 - }, - { - "epoch": 3.8331771321462043, - "grad_norm": 1.3855458498001099, - "learning_rate": 1.7604264292408623e-05, - "loss": 0.1213, - "step": 8180 - }, - { - "epoch": 3.8378631677600747, - "grad_norm": 1.619994044303894, - "learning_rate": 1.7601335520149952e-05, - "loss": 0.1484, - "step": 8190 - }, - { - "epoch": 3.8425492033739457, - "grad_norm": 1.3682477474212646, - "learning_rate": 1.7598406747891285e-05, - "loss": 0.0907, - "step": 8200 - }, - { - "epoch": 3.847235238987816, - "grad_norm": 1.5766955614089966, - "learning_rate": 1.7595477975632615e-05, - "loss": 0.1081, - "step": 8210 - }, - { - "epoch": 3.851921274601687, - "grad_norm": 1.5448287725448608, - "learning_rate": 1.7592549203373948e-05, - "loss": 0.1296, - "step": 8220 - }, - { - "epoch": 3.8566073102155576, - "grad_norm": 1.2130484580993652, - "learning_rate": 1.7589620431115277e-05, - "loss": 0.1141, - "step": 8230 - }, - { - "epoch": 3.861293345829428, - "grad_norm": 1.5601420402526855, - "learning_rate": 1.758669165885661e-05, - "loss": 0.1303, - "step": 8240 - }, - { - "epoch": 3.865979381443299, - "grad_norm": 1.26397705078125, - "learning_rate": 1.758376288659794e-05, - "loss": 0.1017, - "step": 8250 - }, - { - "epoch": 3.865979381443299, - "eval_loss": 0.031061464920639992, - "eval_pearson_cosine": 0.8181771715322625, - "eval_pearson_dot": 0.753218304404399, - "eval_pearson_euclidean": 0.7900298439690836, - "eval_pearson_manhattan": 0.792536151039883, - "eval_runtime": 3.1857, - "eval_samples_per_second": 470.856, - "eval_spearman_cosine": 0.8174158550444287, - "eval_spearman_dot": 0.7522993308222343, - "eval_spearman_euclidean": 0.798562890702385, - "eval_spearman_manhattan": 0.8006887717421057, - "eval_steps_per_second": 29.507, - "step": 8250 - }, - { - "epoch": 3.8706654170571695, - "grad_norm": 1.8911367654800415, - "learning_rate": 1.7580834114339272e-05, - "loss": 0.1197, - "step": 8260 - }, - { - "epoch": 3.8753514526710404, - "grad_norm": 1.5515751838684082, - "learning_rate": 1.7577905342080602e-05, - "loss": 0.1278, - "step": 8270 - }, - { - "epoch": 3.880037488284911, - "grad_norm": 2.1667306423187256, - "learning_rate": 1.757497656982193e-05, - "loss": 0.1387, - "step": 8280 - }, - { - "epoch": 3.8847235238987814, - "grad_norm": 1.149591326713562, - "learning_rate": 1.7572047797563264e-05, - "loss": 0.1231, - "step": 8290 - }, - { - "epoch": 3.8894095595126523, - "grad_norm": 1.848067283630371, - "learning_rate": 1.7569119025304594e-05, - "loss": 0.1228, - "step": 8300 - }, - { - "epoch": 3.894095595126523, - "grad_norm": 1.451674222946167, - "learning_rate": 1.7566190253045923e-05, - "loss": 0.1161, - "step": 8310 - }, - { - "epoch": 3.8987816307403937, - "grad_norm": 1.7280783653259277, - "learning_rate": 1.7563261480787256e-05, - "loss": 0.1191, - "step": 8320 - }, - { - "epoch": 3.9034676663542642, - "grad_norm": 1.5939549207687378, - "learning_rate": 1.7560332708528586e-05, - "loss": 0.1272, - "step": 8330 - }, - { - "epoch": 3.9081537019681347, - "grad_norm": 1.6509348154067993, - "learning_rate": 1.7557403936269915e-05, - "loss": 0.127, - "step": 8340 - }, - { - "epoch": 3.9128397375820057, - "grad_norm": 1.8915349245071411, - "learning_rate": 1.7554475164011248e-05, - "loss": 0.1352, - "step": 8350 - }, - { - "epoch": 3.917525773195876, - "grad_norm": 2.188493490219116, - "learning_rate": 1.7551546391752578e-05, - "loss": 0.1105, - "step": 8360 - }, - { - "epoch": 3.922211808809747, - "grad_norm": 1.8589377403259277, - "learning_rate": 1.754861761949391e-05, - "loss": 0.1031, - "step": 8370 - }, - { - "epoch": 3.9268978444236176, - "grad_norm": 1.7054208517074585, - "learning_rate": 1.754568884723524e-05, - "loss": 0.1165, - "step": 8380 - }, - { - "epoch": 3.931583880037488, - "grad_norm": 1.2826303243637085, - "learning_rate": 1.754276007497657e-05, - "loss": 0.0994, - "step": 8390 - }, - { - "epoch": 3.936269915651359, - "grad_norm": 2.087935209274292, - "learning_rate": 1.7539831302717902e-05, - "loss": 0.1493, - "step": 8400 - }, - { - "epoch": 3.9409559512652295, - "grad_norm": 1.4399867057800293, - "learning_rate": 1.7536902530459232e-05, - "loss": 0.1126, - "step": 8410 - }, - { - "epoch": 3.9456419868791004, - "grad_norm": 2.081295967102051, - "learning_rate": 1.7533973758200565e-05, - "loss": 0.1149, - "step": 8420 - }, - { - "epoch": 3.950328022492971, - "grad_norm": 1.6477272510528564, - "learning_rate": 1.7531044985941894e-05, - "loss": 0.124, - "step": 8430 - }, - { - "epoch": 3.9550140581068414, - "grad_norm": 1.43690025806427, - "learning_rate": 1.7528116213683227e-05, - "loss": 0.1175, - "step": 8440 - }, - { - "epoch": 3.9597000937207123, - "grad_norm": 2.231391429901123, - "learning_rate": 1.7525187441424557e-05, - "loss": 0.1039, - "step": 8450 - }, - { - "epoch": 3.964386129334583, - "grad_norm": 1.699771761894226, - "learning_rate": 1.752225866916589e-05, - "loss": 0.096, - "step": 8460 - }, - { - "epoch": 3.9690721649484537, - "grad_norm": 0.9869770407676697, - "learning_rate": 1.751932989690722e-05, - "loss": 0.1318, - "step": 8470 - }, - { - "epoch": 3.973758200562324, - "grad_norm": 1.2464418411254883, - "learning_rate": 1.751640112464855e-05, - "loss": 0.1025, - "step": 8480 - }, - { - "epoch": 3.9784442361761947, - "grad_norm": 1.8724063634872437, - "learning_rate": 1.751347235238988e-05, - "loss": 0.1154, - "step": 8490 - }, - { - "epoch": 3.9831302717900656, - "grad_norm": 1.6470191478729248, - "learning_rate": 1.751054358013121e-05, - "loss": 0.1132, - "step": 8500 - }, - { - "epoch": 3.9831302717900656, - "eval_loss": 0.03063393384218216, - "eval_pearson_cosine": 0.821106317003462, - "eval_pearson_dot": 0.7578136492444401, - "eval_pearson_euclidean": 0.7882531341441634, - "eval_pearson_manhattan": 0.790852878268538, - "eval_runtime": 3.174, - "eval_samples_per_second": 472.588, - "eval_spearman_cosine": 0.8198241690509209, - "eval_spearman_dot": 0.7578266102334076, - "eval_spearman_euclidean": 0.7967830640080272, - "eval_spearman_manhattan": 0.7991467507473939, - "eval_steps_per_second": 29.616, - "step": 8500 - }, - { - "epoch": 3.987816307403936, - "grad_norm": 1.676721453666687, - "learning_rate": 1.750761480787254e-05, - "loss": 0.1188, - "step": 8510 - }, - { - "epoch": 3.992502343017807, - "grad_norm": 1.187525987625122, - "learning_rate": 1.750468603561387e-05, - "loss": 0.132, - "step": 8520 - }, - { - "epoch": 3.9971883786316775, - "grad_norm": 1.2841644287109375, - "learning_rate": 1.7501757263355203e-05, - "loss": 0.1379, - "step": 8530 } ], "logging_steps": 10, @@ -6536,7 +5369,7 @@ "should_evaluate": false, "should_log": false, "should_save": true, - "should_training_stop": true + "should_training_stop": false }, "attributes": {} }