| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.0, | |
| "eval_steps": 250, | |
| "global_step": 8536, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004686035613870665, | |
| "grad_norm": 0.5978041887283325, | |
| "learning_rate": 1.999707122774133e-05, | |
| "loss": 3.4332, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.00937207122774133, | |
| "grad_norm": 1.5130422115325928, | |
| "learning_rate": 1.9994142455482664e-05, | |
| "loss": 3.2454, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.014058106841611996, | |
| "grad_norm": 1.607094168663025, | |
| "learning_rate": 1.9991213683223994e-05, | |
| "loss": 2.9102, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01874414245548266, | |
| "grad_norm": 2.6665945053100586, | |
| "learning_rate": 1.9988284910965326e-05, | |
| "loss": 2.7066, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.023430178069353328, | |
| "grad_norm": 3.212167978286743, | |
| "learning_rate": 1.9985356138706656e-05, | |
| "loss": 2.5895, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.028116213683223992, | |
| "grad_norm": 3.8364155292510986, | |
| "learning_rate": 1.998242736644799e-05, | |
| "loss": 2.4215, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03280224929709466, | |
| "grad_norm": 3.560389280319214, | |
| "learning_rate": 1.9979498594189318e-05, | |
| "loss": 2.0972, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03748828491096532, | |
| "grad_norm": 4.126266956329346, | |
| "learning_rate": 1.9976569821930648e-05, | |
| "loss": 1.8428, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.04217432052483599, | |
| "grad_norm": 4.506312370300293, | |
| "learning_rate": 1.997364104967198e-05, | |
| "loss": 1.5853, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.046860356138706656, | |
| "grad_norm": 4.499500751495361, | |
| "learning_rate": 1.997071227741331e-05, | |
| "loss": 1.4429, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.05154639175257732, | |
| "grad_norm": 4.746804237365723, | |
| "learning_rate": 1.996778350515464e-05, | |
| "loss": 1.3275, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.056232427366447985, | |
| "grad_norm": 4.36592435836792, | |
| "learning_rate": 1.996485473289597e-05, | |
| "loss": 1.0965, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.06091846298031865, | |
| "grad_norm": 4.177589416503906, | |
| "learning_rate": 1.9961925960637302e-05, | |
| "loss": 1.0654, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.06560449859418932, | |
| "grad_norm": 4.377724647521973, | |
| "learning_rate": 1.995899718837863e-05, | |
| "loss": 0.9687, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.07029053420805999, | |
| "grad_norm": 3.9814295768737793, | |
| "learning_rate": 1.9956068416119964e-05, | |
| "loss": 1.043, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.07497656982193064, | |
| "grad_norm": 3.938041925430298, | |
| "learning_rate": 1.9953139643861294e-05, | |
| "loss": 0.9639, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.07966260543580131, | |
| "grad_norm": 4.323759078979492, | |
| "learning_rate": 1.9950210871602627e-05, | |
| "loss": 0.9344, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.08434864104967198, | |
| "grad_norm": 4.037902355194092, | |
| "learning_rate": 1.9947282099343956e-05, | |
| "loss": 0.8302, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.08903467666354264, | |
| "grad_norm": 3.9691343307495117, | |
| "learning_rate": 1.994435332708529e-05, | |
| "loss": 0.8573, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.09372071227741331, | |
| "grad_norm": 3.8372201919555664, | |
| "learning_rate": 1.994142455482662e-05, | |
| "loss": 0.808, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.09840674789128398, | |
| "grad_norm": 4.224605083465576, | |
| "learning_rate": 1.9938495782567948e-05, | |
| "loss": 0.8942, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.10309278350515463, | |
| "grad_norm": 3.7245891094207764, | |
| "learning_rate": 1.993556701030928e-05, | |
| "loss": 0.8321, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1077788191190253, | |
| "grad_norm": 4.344135761260986, | |
| "learning_rate": 1.993263823805061e-05, | |
| "loss": 0.8035, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.11246485473289597, | |
| "grad_norm": 3.6320858001708984, | |
| "learning_rate": 1.9929709465791943e-05, | |
| "loss": 0.723, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.11715089034676664, | |
| "grad_norm": 2.934122085571289, | |
| "learning_rate": 1.9926780693533273e-05, | |
| "loss": 0.761, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.11715089034676664, | |
| "eval_loss": 0.13966944813728333, | |
| "eval_pearson_cosine": 0.7191493566389795, | |
| "eval_pearson_dot": 0.43422507697410495, | |
| "eval_pearson_euclidean": 0.7135360350736899, | |
| "eval_pearson_manhattan": 0.7129270627309872, | |
| "eval_runtime": 3.3956, | |
| "eval_samples_per_second": 441.749, | |
| "eval_spearman_cosine": 0.7365500082397669, | |
| "eval_spearman_dot": 0.43017280395518565, | |
| "eval_spearman_euclidean": 0.7209662281529704, | |
| "eval_spearman_manhattan": 0.720539010106817, | |
| "eval_steps_per_second": 27.683, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1218369259606373, | |
| "grad_norm": 3.1349332332611084, | |
| "learning_rate": 1.9923851921274602e-05, | |
| "loss": 0.7916, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.12652296157450796, | |
| "grad_norm": 4.0211286544799805, | |
| "learning_rate": 1.9920923149015935e-05, | |
| "loss": 0.7147, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.13120899718837864, | |
| "grad_norm": 3.262589693069458, | |
| "learning_rate": 1.9917994376757265e-05, | |
| "loss": 0.7689, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1358950328022493, | |
| "grad_norm": 4.0925397872924805, | |
| "learning_rate": 1.9915065604498594e-05, | |
| "loss": 0.7556, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.14058106841611998, | |
| "grad_norm": 3.7041923999786377, | |
| "learning_rate": 1.9912136832239924e-05, | |
| "loss": 0.7239, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.14526710402999063, | |
| "grad_norm": 3.712703227996826, | |
| "learning_rate": 1.9909208059981257e-05, | |
| "loss": 0.6685, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.14995313964386128, | |
| "grad_norm": 3.6240854263305664, | |
| "learning_rate": 1.9906279287722586e-05, | |
| "loss": 0.7263, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.15463917525773196, | |
| "grad_norm": 4.013660907745361, | |
| "learning_rate": 1.990335051546392e-05, | |
| "loss": 0.7287, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.15932521087160262, | |
| "grad_norm": 2.8780667781829834, | |
| "learning_rate": 1.990042174320525e-05, | |
| "loss": 0.6709, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.1640112464854733, | |
| "grad_norm": 3.035186290740967, | |
| "learning_rate": 1.989749297094658e-05, | |
| "loss": 0.6977, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.16869728209934395, | |
| "grad_norm": 3.433403491973877, | |
| "learning_rate": 1.989456419868791e-05, | |
| "loss": 0.6656, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.1733833177132146, | |
| "grad_norm": 3.8357644081115723, | |
| "learning_rate": 1.9891635426429244e-05, | |
| "loss": 0.6192, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.1780693533270853, | |
| "grad_norm": 3.506263494491577, | |
| "learning_rate": 1.9888706654170573e-05, | |
| "loss": 0.6516, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.18275538894095594, | |
| "grad_norm": 2.839836835861206, | |
| "learning_rate": 1.9885777881911906e-05, | |
| "loss": 0.6056, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.18744142455482662, | |
| "grad_norm": 3.2694239616394043, | |
| "learning_rate": 1.9882849109653236e-05, | |
| "loss": 0.6753, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.19212746016869728, | |
| "grad_norm": 3.5731680393218994, | |
| "learning_rate": 1.9879920337394565e-05, | |
| "loss": 0.6848, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.19681349578256796, | |
| "grad_norm": 2.8836023807525635, | |
| "learning_rate": 1.9876991565135898e-05, | |
| "loss": 0.659, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.2014995313964386, | |
| "grad_norm": 3.5604217052459717, | |
| "learning_rate": 1.9874062792877228e-05, | |
| "loss": 0.6745, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.20618556701030927, | |
| "grad_norm": 2.7834393978118896, | |
| "learning_rate": 1.9871134020618557e-05, | |
| "loss": 0.623, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.21087160262417995, | |
| "grad_norm": 3.5130224227905273, | |
| "learning_rate": 1.986820524835989e-05, | |
| "loss": 0.579, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.2155576382380506, | |
| "grad_norm": 2.9537723064422607, | |
| "learning_rate": 1.986527647610122e-05, | |
| "loss": 0.6042, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.22024367385192128, | |
| "grad_norm": 2.926415205001831, | |
| "learning_rate": 1.986234770384255e-05, | |
| "loss": 0.5785, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.22492970946579194, | |
| "grad_norm": 4.53303861618042, | |
| "learning_rate": 1.9859418931583882e-05, | |
| "loss": 0.5957, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.2296157450796626, | |
| "grad_norm": 2.5960335731506348, | |
| "learning_rate": 1.985649015932521e-05, | |
| "loss": 0.6223, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.23430178069353327, | |
| "grad_norm": 2.509232997894287, | |
| "learning_rate": 1.985356138706654e-05, | |
| "loss": 0.6275, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.23430178069353327, | |
| "eval_loss": 0.12402218580245972, | |
| "eval_pearson_cosine": 0.7535085617332911, | |
| "eval_pearson_dot": 0.4526570980382658, | |
| "eval_pearson_euclidean": 0.7442016683051946, | |
| "eval_pearson_manhattan": 0.7442320131099294, | |
| "eval_runtime": 3.5327, | |
| "eval_samples_per_second": 424.602, | |
| "eval_spearman_cosine": 0.7638158285362802, | |
| "eval_spearman_dot": 0.4533334110281352, | |
| "eval_spearman_euclidean": 0.7505934110060137, | |
| "eval_spearman_manhattan": 0.7505444222182343, | |
| "eval_steps_per_second": 26.608, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.23898781630740393, | |
| "grad_norm": 3.2064850330352783, | |
| "learning_rate": 1.9850632614807874e-05, | |
| "loss": 0.6257, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.2436738519212746, | |
| "grad_norm": 3.235630512237549, | |
| "learning_rate": 1.9847703842549203e-05, | |
| "loss": 0.6069, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.24835988753514526, | |
| "grad_norm": 2.8067846298217773, | |
| "learning_rate": 1.9844775070290536e-05, | |
| "loss": 0.6137, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.2530459231490159, | |
| "grad_norm": 3.6333608627319336, | |
| "learning_rate": 1.9841846298031866e-05, | |
| "loss": 0.6111, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.25773195876288657, | |
| "grad_norm": 2.410837411880493, | |
| "learning_rate": 1.98389175257732e-05, | |
| "loss": 0.5463, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.2624179943767573, | |
| "grad_norm": 3.0617260932922363, | |
| "learning_rate": 1.9835988753514528e-05, | |
| "loss": 0.5571, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.26710402999062793, | |
| "grad_norm": 2.576580762863159, | |
| "learning_rate": 1.983305998125586e-05, | |
| "loss": 0.5428, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.2717900656044986, | |
| "grad_norm": 3.412797451019287, | |
| "learning_rate": 1.983013120899719e-05, | |
| "loss": 0.5541, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.27647610121836924, | |
| "grad_norm": 2.929755210876465, | |
| "learning_rate": 1.9827202436738523e-05, | |
| "loss": 0.5941, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.28116213683223995, | |
| "grad_norm": 2.80749773979187, | |
| "learning_rate": 1.9824273664479853e-05, | |
| "loss": 0.5702, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.2858481724461106, | |
| "grad_norm": 2.91896390914917, | |
| "learning_rate": 1.9821344892221182e-05, | |
| "loss": 0.5971, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.29053420805998126, | |
| "grad_norm": 2.754328727722168, | |
| "learning_rate": 1.9818416119962512e-05, | |
| "loss": 0.5544, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.2952202436738519, | |
| "grad_norm": 3.4075613021850586, | |
| "learning_rate": 1.9815487347703845e-05, | |
| "loss": 0.5981, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.29990627928772257, | |
| "grad_norm": 2.6344642639160156, | |
| "learning_rate": 1.9812558575445174e-05, | |
| "loss": 0.5539, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.3045923149015933, | |
| "grad_norm": 2.417914867401123, | |
| "learning_rate": 1.9809629803186504e-05, | |
| "loss": 0.5078, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.30927835051546393, | |
| "grad_norm": 2.52506685256958, | |
| "learning_rate": 1.9806701030927837e-05, | |
| "loss": 0.532, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.3139643861293346, | |
| "grad_norm": 2.830923557281494, | |
| "learning_rate": 1.9803772258669166e-05, | |
| "loss": 0.5537, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.31865042174320524, | |
| "grad_norm": 3.1638379096984863, | |
| "learning_rate": 1.98008434864105e-05, | |
| "loss": 0.5375, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.3233364573570759, | |
| "grad_norm": 3.447112798690796, | |
| "learning_rate": 1.979791471415183e-05, | |
| "loss": 0.5755, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.3280224929709466, | |
| "grad_norm": 2.3407933712005615, | |
| "learning_rate": 1.9794985941893158e-05, | |
| "loss": 0.514, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.33270852858481725, | |
| "grad_norm": 3.226551055908203, | |
| "learning_rate": 1.979205716963449e-05, | |
| "loss": 0.5167, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.3373945641986879, | |
| "grad_norm": 3.1462602615356445, | |
| "learning_rate": 1.978912839737582e-05, | |
| "loss": 0.5431, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.34208059981255856, | |
| "grad_norm": 3.051692008972168, | |
| "learning_rate": 1.9786199625117153e-05, | |
| "loss": 0.5129, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.3467666354264292, | |
| "grad_norm": 2.5202085971832275, | |
| "learning_rate": 1.9783270852858483e-05, | |
| "loss": 0.5752, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.3514526710402999, | |
| "grad_norm": 2.6891393661499023, | |
| "learning_rate": 1.9780342080599816e-05, | |
| "loss": 0.5326, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.3514526710402999, | |
| "eval_loss": 0.11494184285402298, | |
| "eval_pearson_cosine": 0.7539853173929743, | |
| "eval_pearson_dot": 0.47863358738842976, | |
| "eval_pearson_euclidean": 0.7326912597509363, | |
| "eval_pearson_manhattan": 0.7320283236539638, | |
| "eval_runtime": 3.5674, | |
| "eval_samples_per_second": 420.478, | |
| "eval_spearman_cosine": 0.7697587773995457, | |
| "eval_spearman_dot": 0.47367550293168154, | |
| "eval_spearman_euclidean": 0.7466430705786936, | |
| "eval_spearman_manhattan": 0.7460682347775467, | |
| "eval_steps_per_second": 26.35, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.3561387066541706, | |
| "grad_norm": 2.9754297733306885, | |
| "learning_rate": 1.9777413308341145e-05, | |
| "loss": 0.4959, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.36082474226804123, | |
| "grad_norm": 3.2427656650543213, | |
| "learning_rate": 1.9774484536082478e-05, | |
| "loss": 0.539, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.3655107778819119, | |
| "grad_norm": 2.6402335166931152, | |
| "learning_rate": 1.9771555763823808e-05, | |
| "loss": 0.5328, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.3701968134957826, | |
| "grad_norm": 2.9217915534973145, | |
| "learning_rate": 1.9768626991565137e-05, | |
| "loss": 0.4895, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.37488284910965325, | |
| "grad_norm": 2.363999128341675, | |
| "learning_rate": 1.976569821930647e-05, | |
| "loss": 0.4805, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.3795688847235239, | |
| "grad_norm": 2.491255044937134, | |
| "learning_rate": 1.97627694470478e-05, | |
| "loss": 0.5281, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.38425492033739456, | |
| "grad_norm": 2.595560312271118, | |
| "learning_rate": 1.975984067478913e-05, | |
| "loss": 0.5104, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.3889409559512652, | |
| "grad_norm": 2.991326332092285, | |
| "learning_rate": 1.975691190253046e-05, | |
| "loss": 0.508, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.3936269915651359, | |
| "grad_norm": 2.6440255641937256, | |
| "learning_rate": 1.975398313027179e-05, | |
| "loss": 0.5643, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.3983130271790066, | |
| "grad_norm": 3.06538987159729, | |
| "learning_rate": 1.975105435801312e-05, | |
| "loss": 0.4944, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.4029990627928772, | |
| "grad_norm": 2.751331329345703, | |
| "learning_rate": 1.9748125585754454e-05, | |
| "loss": 0.5215, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.4076850984067479, | |
| "grad_norm": 3.2387125492095947, | |
| "learning_rate": 1.9745196813495783e-05, | |
| "loss": 0.5925, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.41237113402061853, | |
| "grad_norm": 2.5455057621002197, | |
| "learning_rate": 1.9742268041237116e-05, | |
| "loss": 0.4956, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.41705716963448924, | |
| "grad_norm": 2.864497423171997, | |
| "learning_rate": 1.9739339268978446e-05, | |
| "loss": 0.4855, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.4217432052483599, | |
| "grad_norm": 2.6306967735290527, | |
| "learning_rate": 1.9736410496719775e-05, | |
| "loss": 0.4542, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.42642924086223055, | |
| "grad_norm": 2.294090747833252, | |
| "learning_rate": 1.9733481724461108e-05, | |
| "loss": 0.4247, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.4311152764761012, | |
| "grad_norm": 2.9401843547821045, | |
| "learning_rate": 1.9730552952202438e-05, | |
| "loss": 0.5219, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.43580131208997186, | |
| "grad_norm": 2.5331194400787354, | |
| "learning_rate": 1.972762417994377e-05, | |
| "loss": 0.5193, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.44048734770384257, | |
| "grad_norm": 3.265944719314575, | |
| "learning_rate": 1.97246954076851e-05, | |
| "loss": 0.4733, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.4451733833177132, | |
| "grad_norm": 2.6586129665374756, | |
| "learning_rate": 1.9721766635426433e-05, | |
| "loss": 0.4939, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.4498594189315839, | |
| "grad_norm": 2.126664638519287, | |
| "learning_rate": 1.9718837863167762e-05, | |
| "loss": 0.5149, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.45454545454545453, | |
| "grad_norm": 2.8597347736358643, | |
| "learning_rate": 1.9715909090909092e-05, | |
| "loss": 0.5089, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.4592314901593252, | |
| "grad_norm": 2.5802338123321533, | |
| "learning_rate": 1.9712980318650425e-05, | |
| "loss": 0.5264, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.4639175257731959, | |
| "grad_norm": 2.4247639179229736, | |
| "learning_rate": 1.9710051546391754e-05, | |
| "loss": 0.5416, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.46860356138706655, | |
| "grad_norm": 2.6614298820495605, | |
| "learning_rate": 1.9707122774133084e-05, | |
| "loss": 0.4917, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.46860356138706655, | |
| "eval_loss": 0.1028011292219162, | |
| "eval_pearson_cosine": 0.7630335129215418, | |
| "eval_pearson_dot": 0.5427640329363896, | |
| "eval_pearson_euclidean": 0.7394698860945113, | |
| "eval_pearson_manhattan": 0.7395035409540256, | |
| "eval_runtime": 3.2727, | |
| "eval_samples_per_second": 458.333, | |
| "eval_spearman_cosine": 0.7777749785163052, | |
| "eval_spearman_dot": 0.5403903388556195, | |
| "eval_spearman_euclidean": 0.7530551416243708, | |
| "eval_spearman_manhattan": 0.753178994867466, | |
| "eval_steps_per_second": 28.722, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.4732895970009372, | |
| "grad_norm": 2.8377857208251953, | |
| "learning_rate": 1.9704194001874417e-05, | |
| "loss": 0.4806, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.47797563261480785, | |
| "grad_norm": 2.3730361461639404, | |
| "learning_rate": 1.9701265229615746e-05, | |
| "loss": 0.4946, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.48266166822867856, | |
| "grad_norm": 3.101628541946411, | |
| "learning_rate": 1.9698336457357076e-05, | |
| "loss": 0.4958, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.4873477038425492, | |
| "grad_norm": 2.4200384616851807, | |
| "learning_rate": 1.969540768509841e-05, | |
| "loss": 0.5231, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.49203373945641987, | |
| "grad_norm": 2.9535088539123535, | |
| "learning_rate": 1.9692478912839738e-05, | |
| "loss": 0.4952, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.4967197750702905, | |
| "grad_norm": 2.3175110816955566, | |
| "learning_rate": 1.968955014058107e-05, | |
| "loss": 0.4845, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.5014058106841612, | |
| "grad_norm": 2.934502124786377, | |
| "learning_rate": 1.96866213683224e-05, | |
| "loss": 0.5229, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.5060918462980318, | |
| "grad_norm": 2.5961878299713135, | |
| "learning_rate": 1.9683692596063733e-05, | |
| "loss": 0.5111, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.5107778819119025, | |
| "grad_norm": 2.9093525409698486, | |
| "learning_rate": 1.9680763823805063e-05, | |
| "loss": 0.545, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.5154639175257731, | |
| "grad_norm": 2.733900547027588, | |
| "learning_rate": 1.9677835051546392e-05, | |
| "loss": 0.4589, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.5201499531396439, | |
| "grad_norm": 2.3905482292175293, | |
| "learning_rate": 1.9674906279287725e-05, | |
| "loss": 0.5197, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.5248359887535146, | |
| "grad_norm": 2.5557029247283936, | |
| "learning_rate": 1.9671977507029055e-05, | |
| "loss": 0.5089, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.5295220243673852, | |
| "grad_norm": 2.3789846897125244, | |
| "learning_rate": 1.9669048734770388e-05, | |
| "loss": 0.5373, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.5342080599812559, | |
| "grad_norm": 2.848998785018921, | |
| "learning_rate": 1.9666119962511717e-05, | |
| "loss": 0.5321, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.5388940955951266, | |
| "grad_norm": 2.2063546180725098, | |
| "learning_rate": 1.9663191190253047e-05, | |
| "loss": 0.4809, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.5435801312089972, | |
| "grad_norm": 2.7333438396453857, | |
| "learning_rate": 1.966026241799438e-05, | |
| "loss": 0.4615, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.5482661668228679, | |
| "grad_norm": 3.124687671661377, | |
| "learning_rate": 1.965733364573571e-05, | |
| "loss": 0.5177, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.5529522024367385, | |
| "grad_norm": 2.726681709289551, | |
| "learning_rate": 1.965440487347704e-05, | |
| "loss": 0.4891, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.5576382380506092, | |
| "grad_norm": 2.9802463054656982, | |
| "learning_rate": 1.965147610121837e-05, | |
| "loss": 0.5018, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.5623242736644799, | |
| "grad_norm": 2.844006299972534, | |
| "learning_rate": 1.96485473289597e-05, | |
| "loss": 0.5149, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.5670103092783505, | |
| "grad_norm": 2.442472219467163, | |
| "learning_rate": 1.964561855670103e-05, | |
| "loss": 0.4808, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.5716963448922212, | |
| "grad_norm": 1.9972388744354248, | |
| "learning_rate": 1.9642689784442363e-05, | |
| "loss": 0.4743, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.5763823805060918, | |
| "grad_norm": 2.6430776119232178, | |
| "learning_rate": 1.9639761012183693e-05, | |
| "loss": 0.5057, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.5810684161199625, | |
| "grad_norm": 2.37652850151062, | |
| "learning_rate": 1.9636832239925026e-05, | |
| "loss": 0.4851, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.5857544517338332, | |
| "grad_norm": 2.755315065383911, | |
| "learning_rate": 1.9633903467666355e-05, | |
| "loss": 0.4451, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.5857544517338332, | |
| "eval_loss": 0.09590280055999756, | |
| "eval_pearson_cosine": 0.7634019154373535, | |
| "eval_pearson_dot": 0.5909367397822605, | |
| "eval_pearson_euclidean": 0.7507597414427352, | |
| "eval_pearson_manhattan": 0.7505103943331619, | |
| "eval_runtime": 3.3916, | |
| "eval_samples_per_second": 442.267, | |
| "eval_spearman_cosine": 0.7803407693262222, | |
| "eval_spearman_dot": 0.592884901593322, | |
| "eval_spearman_euclidean": 0.7652017734526011, | |
| "eval_spearman_manhattan": 0.7648713625602643, | |
| "eval_steps_per_second": 27.715, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.5904404873477038, | |
| "grad_norm": 3.0767931938171387, | |
| "learning_rate": 1.9630974695407688e-05, | |
| "loss": 0.5093, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.5951265229615745, | |
| "grad_norm": 3.0984067916870117, | |
| "learning_rate": 1.9628045923149017e-05, | |
| "loss": 0.4888, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.5998125585754451, | |
| "grad_norm": 2.496957778930664, | |
| "learning_rate": 1.962511715089035e-05, | |
| "loss": 0.5005, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.6044985941893158, | |
| "grad_norm": 2.1109137535095215, | |
| "learning_rate": 1.962218837863168e-05, | |
| "loss": 0.4729, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.6091846298031866, | |
| "grad_norm": 2.663804054260254, | |
| "learning_rate": 1.961925960637301e-05, | |
| "loss": 0.4454, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.6138706654170571, | |
| "grad_norm": 1.8708107471466064, | |
| "learning_rate": 1.9616330834114342e-05, | |
| "loss": 0.4949, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.6185567010309279, | |
| "grad_norm": 2.9178926944732666, | |
| "learning_rate": 1.9613402061855672e-05, | |
| "loss": 0.4438, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.6232427366447985, | |
| "grad_norm": 2.1928579807281494, | |
| "learning_rate": 1.9610473289597005e-05, | |
| "loss": 0.4483, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.6279287722586692, | |
| "grad_norm": 2.3122403621673584, | |
| "learning_rate": 1.9607544517338334e-05, | |
| "loss": 0.4677, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.6326148078725399, | |
| "grad_norm": 2.5242831707000732, | |
| "learning_rate": 1.9604615745079664e-05, | |
| "loss": 0.4822, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.6373008434864105, | |
| "grad_norm": 2.572446346282959, | |
| "learning_rate": 1.9601686972820993e-05, | |
| "loss": 0.46, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.6419868791002812, | |
| "grad_norm": 2.6114673614501953, | |
| "learning_rate": 1.9598758200562326e-05, | |
| "loss": 0.4606, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.6466729147141518, | |
| "grad_norm": 2.6151556968688965, | |
| "learning_rate": 1.9595829428303655e-05, | |
| "loss": 0.459, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.6513589503280225, | |
| "grad_norm": 2.9040744304656982, | |
| "learning_rate": 1.959290065604499e-05, | |
| "loss": 0.4994, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.6560449859418932, | |
| "grad_norm": 2.9549591541290283, | |
| "learning_rate": 1.9589971883786318e-05, | |
| "loss": 0.4335, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.6607310215557638, | |
| "grad_norm": 2.2790753841400146, | |
| "learning_rate": 1.9587043111527647e-05, | |
| "loss": 0.4642, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.6654170571696345, | |
| "grad_norm": 2.265308141708374, | |
| "learning_rate": 1.958411433926898e-05, | |
| "loss": 0.4654, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.6701030927835051, | |
| "grad_norm": 2.1528425216674805, | |
| "learning_rate": 1.958118556701031e-05, | |
| "loss": 0.4983, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.6747891283973758, | |
| "grad_norm": 2.840533494949341, | |
| "learning_rate": 1.9578256794751643e-05, | |
| "loss": 0.4989, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.6794751640112465, | |
| "grad_norm": 2.4118032455444336, | |
| "learning_rate": 1.9575328022492972e-05, | |
| "loss": 0.4207, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.6841611996251171, | |
| "grad_norm": 2.845712900161743, | |
| "learning_rate": 1.9572399250234305e-05, | |
| "loss": 0.4219, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.6888472352389878, | |
| "grad_norm": 2.6068544387817383, | |
| "learning_rate": 1.9569470477975635e-05, | |
| "loss": 0.4618, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.6935332708528584, | |
| "grad_norm": 2.6239466667175293, | |
| "learning_rate": 1.9566541705716967e-05, | |
| "loss": 0.4613, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.6982193064667291, | |
| "grad_norm": 2.417541742324829, | |
| "learning_rate": 1.9563612933458297e-05, | |
| "loss": 0.4457, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.7029053420805998, | |
| "grad_norm": 2.444243907928467, | |
| "learning_rate": 1.9560684161199626e-05, | |
| "loss": 0.4682, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.7029053420805998, | |
| "eval_loss": 0.10573244839906693, | |
| "eval_pearson_cosine": 0.7687240391174335, | |
| "eval_pearson_dot": 0.52706539661569, | |
| "eval_pearson_euclidean": 0.7544794171981408, | |
| "eval_pearson_manhattan": 0.7540749762273435, | |
| "eval_runtime": 3.5044, | |
| "eval_samples_per_second": 428.028, | |
| "eval_spearman_cosine": 0.7854868415020917, | |
| "eval_spearman_dot": 0.5189687789236848, | |
| "eval_spearman_euclidean": 0.7685180176151998, | |
| "eval_spearman_manhattan": 0.7680518081256867, | |
| "eval_steps_per_second": 26.823, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.7075913776944704, | |
| "grad_norm": 2.4107391834259033, | |
| "learning_rate": 1.955775538894096e-05, | |
| "loss": 0.4231, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.7122774133083412, | |
| "grad_norm": 2.2897465229034424, | |
| "learning_rate": 1.955482661668229e-05, | |
| "loss": 0.495, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.7169634489222118, | |
| "grad_norm": 2.1320059299468994, | |
| "learning_rate": 1.9551897844423618e-05, | |
| "loss": 0.4362, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.7216494845360825, | |
| "grad_norm": 2.1766531467437744, | |
| "learning_rate": 1.9548969072164948e-05, | |
| "loss": 0.4324, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.7263355201499532, | |
| "grad_norm": 2.1343445777893066, | |
| "learning_rate": 1.954604029990628e-05, | |
| "loss": 0.4356, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.7310215557638238, | |
| "grad_norm": 2.6357688903808594, | |
| "learning_rate": 1.954311152764761e-05, | |
| "loss": 0.4617, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.7357075913776945, | |
| "grad_norm": 2.2863566875457764, | |
| "learning_rate": 1.9540182755388943e-05, | |
| "loss": 0.4444, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.7403936269915652, | |
| "grad_norm": 1.9845598936080933, | |
| "learning_rate": 1.9537253983130273e-05, | |
| "loss": 0.4459, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.7450796626054358, | |
| "grad_norm": 2.3808443546295166, | |
| "learning_rate": 1.9534325210871605e-05, | |
| "loss": 0.4664, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.7497656982193065, | |
| "grad_norm": 2.976896286010742, | |
| "learning_rate": 1.9531396438612935e-05, | |
| "loss": 0.4317, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.7544517338331771, | |
| "grad_norm": 3.114450693130493, | |
| "learning_rate": 1.9528467666354264e-05, | |
| "loss": 0.4365, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.7591377694470478, | |
| "grad_norm": 2.208798408508301, | |
| "learning_rate": 1.9525538894095597e-05, | |
| "loss": 0.4357, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.7638238050609185, | |
| "grad_norm": 3.005563259124756, | |
| "learning_rate": 1.9522610121836927e-05, | |
| "loss": 0.4184, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.7685098406747891, | |
| "grad_norm": 2.711338758468628, | |
| "learning_rate": 1.951968134957826e-05, | |
| "loss": 0.4479, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.7731958762886598, | |
| "grad_norm": 1.9399685859680176, | |
| "learning_rate": 1.951675257731959e-05, | |
| "loss": 0.4642, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.7778819119025304, | |
| "grad_norm": 2.5553548336029053, | |
| "learning_rate": 1.9513823805060922e-05, | |
| "loss": 0.4851, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.7825679475164011, | |
| "grad_norm": 2.7015600204467773, | |
| "learning_rate": 1.951089503280225e-05, | |
| "loss": 0.4294, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.7872539831302718, | |
| "grad_norm": 3.027247428894043, | |
| "learning_rate": 1.950796626054358e-05, | |
| "loss": 0.4359, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.7919400187441424, | |
| "grad_norm": 2.603200674057007, | |
| "learning_rate": 1.9505037488284914e-05, | |
| "loss": 0.4833, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.7966260543580131, | |
| "grad_norm": 2.2245004177093506, | |
| "learning_rate": 1.9502108716026243e-05, | |
| "loss": 0.4633, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.8013120899718837, | |
| "grad_norm": 2.1866180896759033, | |
| "learning_rate": 1.9499179943767573e-05, | |
| "loss": 0.4465, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.8059981255857545, | |
| "grad_norm": 2.464083671569824, | |
| "learning_rate": 1.9496251171508906e-05, | |
| "loss": 0.4282, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.8106841611996252, | |
| "grad_norm": 2.5016162395477295, | |
| "learning_rate": 1.9493322399250235e-05, | |
| "loss": 0.4566, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.8153701968134958, | |
| "grad_norm": 2.9877169132232666, | |
| "learning_rate": 1.9490393626991565e-05, | |
| "loss": 0.462, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.8200562324273665, | |
| "grad_norm": 2.2195188999176025, | |
| "learning_rate": 1.9487464854732898e-05, | |
| "loss": 0.4489, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.8200562324273665, | |
| "eval_loss": 0.09936515241861343, | |
| "eval_pearson_cosine": 0.7657869832789288, | |
| "eval_pearson_dot": 0.576463122226734, | |
| "eval_pearson_euclidean": 0.7513952467595857, | |
| "eval_pearson_manhattan": 0.7505485236956133, | |
| "eval_runtime": 3.3329, | |
| "eval_samples_per_second": 450.052, | |
| "eval_spearman_cosine": 0.7800032031654975, | |
| "eval_spearman_dot": 0.576004160544721, | |
| "eval_spearman_euclidean": 0.7627214415950069, | |
| "eval_spearman_manhattan": 0.7623633881575729, | |
| "eval_steps_per_second": 28.203, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.8247422680412371, | |
| "grad_norm": 2.793635606765747, | |
| "learning_rate": 1.9484536082474227e-05, | |
| "loss": 0.4434, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.8294283036551078, | |
| "grad_norm": 2.403629779815674, | |
| "learning_rate": 1.948160731021556e-05, | |
| "loss": 0.4034, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.8341143392689785, | |
| "grad_norm": 2.6248741149902344, | |
| "learning_rate": 1.947867853795689e-05, | |
| "loss": 0.4328, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.8388003748828491, | |
| "grad_norm": 1.8423967361450195, | |
| "learning_rate": 1.9475749765698223e-05, | |
| "loss": 0.458, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.8434864104967198, | |
| "grad_norm": 2.8771181106567383, | |
| "learning_rate": 1.9472820993439552e-05, | |
| "loss": 0.3796, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.8481724461105904, | |
| "grad_norm": 2.6302261352539062, | |
| "learning_rate": 1.946989222118088e-05, | |
| "loss": 0.4494, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.8528584817244611, | |
| "grad_norm": 1.8728386163711548, | |
| "learning_rate": 1.9466963448922214e-05, | |
| "loss": 0.4624, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.8575445173383318, | |
| "grad_norm": 1.9615451097488403, | |
| "learning_rate": 1.9464034676663544e-05, | |
| "loss": 0.4666, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.8622305529522024, | |
| "grad_norm": 2.423076629638672, | |
| "learning_rate": 1.9461105904404877e-05, | |
| "loss": 0.4659, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.8669165885660731, | |
| "grad_norm": 2.1707210540771484, | |
| "learning_rate": 1.9458177132146206e-05, | |
| "loss": 0.4536, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.8716026241799437, | |
| "grad_norm": 2.3241922855377197, | |
| "learning_rate": 1.9455248359887536e-05, | |
| "loss": 0.4589, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.8762886597938144, | |
| "grad_norm": 2.628664255142212, | |
| "learning_rate": 1.945231958762887e-05, | |
| "loss": 0.4801, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.8809746954076851, | |
| "grad_norm": 2.25126314163208, | |
| "learning_rate": 1.9449390815370198e-05, | |
| "loss": 0.4127, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.8856607310215557, | |
| "grad_norm": 2.988956928253174, | |
| "learning_rate": 1.9446462043111528e-05, | |
| "loss": 0.4431, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.8903467666354264, | |
| "grad_norm": 2.3071420192718506, | |
| "learning_rate": 1.944353327085286e-05, | |
| "loss": 0.4606, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.895032802249297, | |
| "grad_norm": 2.712460517883301, | |
| "learning_rate": 1.944060449859419e-05, | |
| "loss": 0.4853, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.8997188378631678, | |
| "grad_norm": 2.555626153945923, | |
| "learning_rate": 1.943767572633552e-05, | |
| "loss": 0.4187, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.9044048734770385, | |
| "grad_norm": 2.3509678840637207, | |
| "learning_rate": 1.9434746954076852e-05, | |
| "loss": 0.4144, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.9090909090909091, | |
| "grad_norm": 2.63004994392395, | |
| "learning_rate": 1.9431818181818182e-05, | |
| "loss": 0.399, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.9137769447047798, | |
| "grad_norm": 2.8148534297943115, | |
| "learning_rate": 1.9428889409559515e-05, | |
| "loss": 0.4309, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.9184629803186504, | |
| "grad_norm": 2.594778537750244, | |
| "learning_rate": 1.9425960637300844e-05, | |
| "loss": 0.4753, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.9231490159325211, | |
| "grad_norm": 2.4502975940704346, | |
| "learning_rate": 1.9423031865042177e-05, | |
| "loss": 0.4317, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.9278350515463918, | |
| "grad_norm": 2.335707664489746, | |
| "learning_rate": 1.9420103092783507e-05, | |
| "loss": 0.405, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.9325210871602624, | |
| "grad_norm": 2.0419764518737793, | |
| "learning_rate": 1.941717432052484e-05, | |
| "loss": 0.4669, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.9372071227741331, | |
| "grad_norm": 2.2041420936584473, | |
| "learning_rate": 1.941424554826617e-05, | |
| "loss": 0.4696, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.9372071227741331, | |
| "eval_loss": 0.10549145191907883, | |
| "eval_pearson_cosine": 0.7617707450696116, | |
| "eval_pearson_dot": 0.5909647410786931, | |
| "eval_pearson_euclidean": 0.7526201710813609, | |
| "eval_pearson_manhattan": 0.7514050722082715, | |
| "eval_runtime": 3.4394, | |
| "eval_samples_per_second": 436.124, | |
| "eval_spearman_cosine": 0.7834635716420068, | |
| "eval_spearman_dot": 0.5835376546734291, | |
| "eval_spearman_euclidean": 0.7675113486625056, | |
| "eval_spearman_manhattan": 0.7668717701506194, | |
| "eval_steps_per_second": 27.33, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.9418931583880038, | |
| "grad_norm": 1.7690527439117432, | |
| "learning_rate": 1.94113167760075e-05, | |
| "loss": 0.39, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.9465791940018744, | |
| "grad_norm": 2.247870683670044, | |
| "learning_rate": 1.940838800374883e-05, | |
| "loss": 0.4283, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.9512652296157451, | |
| "grad_norm": 2.7865118980407715, | |
| "learning_rate": 1.940545923149016e-05, | |
| "loss": 0.4041, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.9559512652296157, | |
| "grad_norm": 2.7173750400543213, | |
| "learning_rate": 1.9402530459231494e-05, | |
| "loss": 0.428, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.9606373008434864, | |
| "grad_norm": 2.4290664196014404, | |
| "learning_rate": 1.9399601686972823e-05, | |
| "loss": 0.4363, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.9653233364573571, | |
| "grad_norm": 2.9478437900543213, | |
| "learning_rate": 1.9396672914714153e-05, | |
| "loss": 0.4341, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.9700093720712277, | |
| "grad_norm": 2.5642406940460205, | |
| "learning_rate": 1.9393744142455482e-05, | |
| "loss": 0.4402, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.9746954076850984, | |
| "grad_norm": 2.622969388961792, | |
| "learning_rate": 1.9390815370196815e-05, | |
| "loss": 0.4415, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.979381443298969, | |
| "grad_norm": 1.8864187002182007, | |
| "learning_rate": 1.9387886597938145e-05, | |
| "loss": 0.421, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.9840674789128397, | |
| "grad_norm": 2.145385503768921, | |
| "learning_rate": 1.9384957825679474e-05, | |
| "loss": 0.4142, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.9887535145267105, | |
| "grad_norm": 2.5870542526245117, | |
| "learning_rate": 1.9382029053420807e-05, | |
| "loss": 0.4701, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.993439550140581, | |
| "grad_norm": 2.5176923274993896, | |
| "learning_rate": 1.9379100281162137e-05, | |
| "loss": 0.3776, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.9981255857544518, | |
| "grad_norm": 2.8260464668273926, | |
| "learning_rate": 1.937617150890347e-05, | |
| "loss": 0.4453, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.0028116213683225, | |
| "grad_norm": 2.2735579013824463, | |
| "learning_rate": 1.93732427366448e-05, | |
| "loss": 0.3617, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.007497656982193, | |
| "grad_norm": 1.9802652597427368, | |
| "learning_rate": 1.9370313964386132e-05, | |
| "loss": 0.3103, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.0121836925960637, | |
| "grad_norm": 2.4149436950683594, | |
| "learning_rate": 1.936738519212746e-05, | |
| "loss": 0.3183, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.0168697282099344, | |
| "grad_norm": 2.9166879653930664, | |
| "learning_rate": 1.9364456419868794e-05, | |
| "loss": 0.2927, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.021555763823805, | |
| "grad_norm": 1.8774973154067993, | |
| "learning_rate": 1.9361527647610124e-05, | |
| "loss": 0.3217, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.0262417994376758, | |
| "grad_norm": 2.0291717052459717, | |
| "learning_rate": 1.9358598875351457e-05, | |
| "loss": 0.3236, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.0309278350515463, | |
| "grad_norm": 2.500614643096924, | |
| "learning_rate": 1.9355670103092786e-05, | |
| "loss": 0.3187, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.035613870665417, | |
| "grad_norm": 2.4826183319091797, | |
| "learning_rate": 1.9352741330834116e-05, | |
| "loss": 0.3284, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.0402999062792877, | |
| "grad_norm": 2.647724151611328, | |
| "learning_rate": 1.934981255857545e-05, | |
| "loss": 0.3432, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.0449859418931584, | |
| "grad_norm": 2.356365203857422, | |
| "learning_rate": 1.9346883786316778e-05, | |
| "loss": 0.3321, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.0496719775070291, | |
| "grad_norm": 2.514002561569214, | |
| "learning_rate": 1.9343955014058108e-05, | |
| "loss": 0.3115, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.0543580131208996, | |
| "grad_norm": 2.049299716949463, | |
| "learning_rate": 1.9341026241799437e-05, | |
| "loss": 0.3474, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.0543580131208996, | |
| "eval_loss": 0.0818348377943039, | |
| "eval_pearson_cosine": 0.7662947284133068, | |
| "eval_pearson_dot": 0.5774155023095773, | |
| "eval_pearson_euclidean": 0.7536499802604126, | |
| "eval_pearson_manhattan": 0.7526750534069606, | |
| "eval_runtime": 3.4518, | |
| "eval_samples_per_second": 434.553, | |
| "eval_spearman_cosine": 0.7776893657594519, | |
| "eval_spearman_dot": 0.5747770364870878, | |
| "eval_spearman_euclidean": 0.7641629894500831, | |
| "eval_spearman_manhattan": 0.7635836296791149, | |
| "eval_steps_per_second": 27.232, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.0590440487347703, | |
| "grad_norm": 1.9242613315582275, | |
| "learning_rate": 1.933809746954077e-05, | |
| "loss": 0.3362, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.063730084348641, | |
| "grad_norm": 2.6948235034942627, | |
| "learning_rate": 1.93351686972821e-05, | |
| "loss": 0.3524, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.0684161199625117, | |
| "grad_norm": 2.162079334259033, | |
| "learning_rate": 1.9332239925023432e-05, | |
| "loss": 0.3292, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.0731021555763824, | |
| "grad_norm": 2.1351771354675293, | |
| "learning_rate": 1.9329311152764762e-05, | |
| "loss": 0.3409, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.077788191190253, | |
| "grad_norm": 2.527005434036255, | |
| "learning_rate": 1.932638238050609e-05, | |
| "loss": 0.3027, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.0824742268041236, | |
| "grad_norm": 2.446911334991455, | |
| "learning_rate": 1.9323453608247424e-05, | |
| "loss": 0.3376, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.0871602624179943, | |
| "grad_norm": 2.478301763534546, | |
| "learning_rate": 1.9320524835988754e-05, | |
| "loss": 0.298, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.091846298031865, | |
| "grad_norm": 2.537336587905884, | |
| "learning_rate": 1.9317596063730087e-05, | |
| "loss": 0.3312, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 1.0965323336457358, | |
| "grad_norm": 2.4405462741851807, | |
| "learning_rate": 1.9314667291471416e-05, | |
| "loss": 0.2998, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.1012183692596063, | |
| "grad_norm": 2.2255520820617676, | |
| "learning_rate": 1.931173851921275e-05, | |
| "loss": 0.2987, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.105904404873477, | |
| "grad_norm": 2.541879177093506, | |
| "learning_rate": 1.930880974695408e-05, | |
| "loss": 0.3286, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.1105904404873477, | |
| "grad_norm": 2.1010048389434814, | |
| "learning_rate": 1.930588097469541e-05, | |
| "loss": 0.2989, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 1.1152764761012184, | |
| "grad_norm": 1.9881564378738403, | |
| "learning_rate": 1.930295220243674e-05, | |
| "loss": 0.3537, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.119962511715089, | |
| "grad_norm": 2.503845691680908, | |
| "learning_rate": 1.930002343017807e-05, | |
| "loss": 0.3242, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 1.1246485473289598, | |
| "grad_norm": 2.773836612701416, | |
| "learning_rate": 1.9297094657919403e-05, | |
| "loss": 0.3144, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.1293345829428303, | |
| "grad_norm": 2.7757749557495117, | |
| "learning_rate": 1.9294165885660733e-05, | |
| "loss": 0.3114, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 1.134020618556701, | |
| "grad_norm": 2.707315444946289, | |
| "learning_rate": 1.9291237113402062e-05, | |
| "loss": 0.2872, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.1387066541705717, | |
| "grad_norm": 2.1421258449554443, | |
| "learning_rate": 1.9288308341143395e-05, | |
| "loss": 0.3122, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 1.1433926897844424, | |
| "grad_norm": 2.2792320251464844, | |
| "learning_rate": 1.9285379568884725e-05, | |
| "loss": 0.304, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.148078725398313, | |
| "grad_norm": 2.1971569061279297, | |
| "learning_rate": 1.9282450796626054e-05, | |
| "loss": 0.3571, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 1.1527647610121836, | |
| "grad_norm": 2.101923704147339, | |
| "learning_rate": 1.9279522024367387e-05, | |
| "loss": 0.3069, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.1574507966260543, | |
| "grad_norm": 2.395801305770874, | |
| "learning_rate": 1.9276593252108717e-05, | |
| "loss": 0.3145, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 1.162136832239925, | |
| "grad_norm": 1.9261037111282349, | |
| "learning_rate": 1.927366447985005e-05, | |
| "loss": 0.3229, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.1668228678537957, | |
| "grad_norm": 2.185063362121582, | |
| "learning_rate": 1.927073570759138e-05, | |
| "loss": 0.3289, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 1.1715089034676662, | |
| "grad_norm": 2.2372336387634277, | |
| "learning_rate": 1.926780693533271e-05, | |
| "loss": 0.319, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.1715089034676662, | |
| "eval_loss": 0.07521134614944458, | |
| "eval_pearson_cosine": 0.7753374743054451, | |
| "eval_pearson_dot": 0.5929128805152892, | |
| "eval_pearson_euclidean": 0.7592332760910097, | |
| "eval_pearson_manhattan": 0.7589109467518398, | |
| "eval_runtime": 3.6686, | |
| "eval_samples_per_second": 408.877, | |
| "eval_spearman_cosine": 0.785757446712241, | |
| "eval_spearman_dot": 0.5918750155733082, | |
| "eval_spearman_euclidean": 0.7692376569048597, | |
| "eval_spearman_manhattan": 0.7692457770843739, | |
| "eval_steps_per_second": 25.623, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.176194939081537, | |
| "grad_norm": 2.2003061771392822, | |
| "learning_rate": 1.926487816307404e-05, | |
| "loss": 0.2979, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 1.1808809746954076, | |
| "grad_norm": 1.9160058498382568, | |
| "learning_rate": 1.926194939081537e-05, | |
| "loss": 0.3119, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.1855670103092784, | |
| "grad_norm": 2.5396270751953125, | |
| "learning_rate": 1.9259020618556704e-05, | |
| "loss": 0.2882, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 1.190253045923149, | |
| "grad_norm": 2.184056520462036, | |
| "learning_rate": 1.9256091846298033e-05, | |
| "loss": 0.3012, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.1949390815370198, | |
| "grad_norm": 2.4930028915405273, | |
| "learning_rate": 1.9253163074039366e-05, | |
| "loss": 0.3399, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 1.1996251171508903, | |
| "grad_norm": 2.0577898025512695, | |
| "learning_rate": 1.9250234301780696e-05, | |
| "loss": 0.3013, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.204311152764761, | |
| "grad_norm": 1.9735450744628906, | |
| "learning_rate": 1.9247305529522025e-05, | |
| "loss": 0.315, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 1.2089971883786317, | |
| "grad_norm": 2.815910577774048, | |
| "learning_rate": 1.9244376757263358e-05, | |
| "loss": 0.3134, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.2136832239925024, | |
| "grad_norm": 2.4166650772094727, | |
| "learning_rate": 1.9241447985004687e-05, | |
| "loss": 0.2924, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 1.218369259606373, | |
| "grad_norm": 2.65451979637146, | |
| "learning_rate": 1.9238519212746017e-05, | |
| "loss": 0.313, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.2230552952202436, | |
| "grad_norm": 3.048393726348877, | |
| "learning_rate": 1.923559044048735e-05, | |
| "loss": 0.3396, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 1.2277413308341143, | |
| "grad_norm": 2.1255080699920654, | |
| "learning_rate": 1.923266166822868e-05, | |
| "loss": 0.2918, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.232427366447985, | |
| "grad_norm": 2.4257168769836426, | |
| "learning_rate": 1.922973289597001e-05, | |
| "loss": 0.3221, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 1.2371134020618557, | |
| "grad_norm": 2.3977224826812744, | |
| "learning_rate": 1.9226804123711342e-05, | |
| "loss": 0.3777, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.2417994376757264, | |
| "grad_norm": 2.0743911266326904, | |
| "learning_rate": 1.922387535145267e-05, | |
| "loss": 0.2981, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 1.246485473289597, | |
| "grad_norm": 2.740622043609619, | |
| "learning_rate": 1.9220946579194004e-05, | |
| "loss": 0.3239, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.2511715089034676, | |
| "grad_norm": 2.5174942016601562, | |
| "learning_rate": 1.9218017806935334e-05, | |
| "loss": 0.2981, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 1.2558575445173383, | |
| "grad_norm": 2.3390488624572754, | |
| "learning_rate": 1.9215089034676667e-05, | |
| "loss": 0.3509, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.260543580131209, | |
| "grad_norm": 2.495445966720581, | |
| "learning_rate": 1.9212160262417996e-05, | |
| "loss": 0.3153, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 1.2652296157450795, | |
| "grad_norm": 2.384673595428467, | |
| "learning_rate": 1.9209231490159326e-05, | |
| "loss": 0.3541, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.2699156513589505, | |
| "grad_norm": 1.9707324504852295, | |
| "learning_rate": 1.920630271790066e-05, | |
| "loss": 0.3053, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 1.274601686972821, | |
| "grad_norm": 2.0584781169891357, | |
| "learning_rate": 1.9203373945641988e-05, | |
| "loss": 0.2716, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.2792877225866917, | |
| "grad_norm": 1.9361029863357544, | |
| "learning_rate": 1.920044517338332e-05, | |
| "loss": 0.2988, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 1.2839737582005624, | |
| "grad_norm": 2.3929100036621094, | |
| "learning_rate": 1.919751640112465e-05, | |
| "loss": 0.3356, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.2886597938144329, | |
| "grad_norm": 2.7867496013641357, | |
| "learning_rate": 1.9194587628865983e-05, | |
| "loss": 0.3682, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.2886597938144329, | |
| "eval_loss": 0.07667941600084305, | |
| "eval_pearson_cosine": 0.7735852714798206, | |
| "eval_pearson_dot": 0.578414803552846, | |
| "eval_pearson_euclidean": 0.7564370361137946, | |
| "eval_pearson_manhattan": 0.7556229015098381, | |
| "eval_runtime": 3.3395, | |
| "eval_samples_per_second": 449.165, | |
| "eval_spearman_cosine": 0.7851408708188664, | |
| "eval_spearman_dot": 0.5785312770170487, | |
| "eval_spearman_euclidean": 0.7670951414033645, | |
| "eval_spearman_manhattan": 0.7667469763884154, | |
| "eval_steps_per_second": 28.148, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.2933458294283038, | |
| "grad_norm": 2.0457334518432617, | |
| "learning_rate": 1.9191658856607313e-05, | |
| "loss": 0.3118, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.2980318650421743, | |
| "grad_norm": 2.1723532676696777, | |
| "learning_rate": 1.9188730084348642e-05, | |
| "loss": 0.2918, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 1.302717900656045, | |
| "grad_norm": 2.041738748550415, | |
| "learning_rate": 1.9185801312089972e-05, | |
| "loss": 0.3084, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.3074039362699157, | |
| "grad_norm": 2.774698257446289, | |
| "learning_rate": 1.9182872539831305e-05, | |
| "loss": 0.3337, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 1.3120899718837864, | |
| "grad_norm": 1.9762837886810303, | |
| "learning_rate": 1.9179943767572634e-05, | |
| "loss": 0.3029, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.316776007497657, | |
| "grad_norm": 2.1785385608673096, | |
| "learning_rate": 1.9177014995313964e-05, | |
| "loss": 0.3337, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 1.3214620431115276, | |
| "grad_norm": 2.0208494663238525, | |
| "learning_rate": 1.9174086223055296e-05, | |
| "loss": 0.2931, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.3261480787253983, | |
| "grad_norm": 2.3976972103118896, | |
| "learning_rate": 1.9171157450796626e-05, | |
| "loss": 0.3316, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 1.330834114339269, | |
| "grad_norm": 1.9782285690307617, | |
| "learning_rate": 1.916822867853796e-05, | |
| "loss": 0.3173, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.3355201499531397, | |
| "grad_norm": 2.0938830375671387, | |
| "learning_rate": 1.916529990627929e-05, | |
| "loss": 0.2858, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.3402061855670104, | |
| "grad_norm": 2.079247236251831, | |
| "learning_rate": 1.916237113402062e-05, | |
| "loss": 0.2911, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.344892221180881, | |
| "grad_norm": 2.5487163066864014, | |
| "learning_rate": 1.915944236176195e-05, | |
| "loss": 0.2782, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 1.3495782567947516, | |
| "grad_norm": 2.4930672645568848, | |
| "learning_rate": 1.9156513589503284e-05, | |
| "loss": 0.309, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.3542642924086223, | |
| "grad_norm": 2.208987236022949, | |
| "learning_rate": 1.9153584817244613e-05, | |
| "loss": 0.302, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 1.358950328022493, | |
| "grad_norm": 2.2777154445648193, | |
| "learning_rate": 1.9150656044985943e-05, | |
| "loss": 0.2932, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.3636363636363638, | |
| "grad_norm": 2.0704426765441895, | |
| "learning_rate": 1.9147727272727276e-05, | |
| "loss": 0.3326, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 1.3683223992502342, | |
| "grad_norm": 2.550323486328125, | |
| "learning_rate": 1.9144798500468605e-05, | |
| "loss": 0.2945, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.373008434864105, | |
| "grad_norm": 2.1471896171569824, | |
| "learning_rate": 1.9141869728209938e-05, | |
| "loss": 0.3258, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 1.3776944704779757, | |
| "grad_norm": 2.4434566497802734, | |
| "learning_rate": 1.9138940955951267e-05, | |
| "loss": 0.3, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.3823805060918464, | |
| "grad_norm": 2.2125484943389893, | |
| "learning_rate": 1.9136012183692597e-05, | |
| "loss": 0.3618, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 1.387066541705717, | |
| "grad_norm": 2.5156352519989014, | |
| "learning_rate": 1.913308341143393e-05, | |
| "loss": 0.3102, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.3917525773195876, | |
| "grad_norm": 2.1281540393829346, | |
| "learning_rate": 1.913015463917526e-05, | |
| "loss": 0.31, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 1.3964386129334583, | |
| "grad_norm": 2.1260693073272705, | |
| "learning_rate": 1.912722586691659e-05, | |
| "loss": 0.2946, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.401124648547329, | |
| "grad_norm": 2.591770887374878, | |
| "learning_rate": 1.912429709465792e-05, | |
| "loss": 0.2906, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 1.4058106841611997, | |
| "grad_norm": 1.832505226135254, | |
| "learning_rate": 1.912136832239925e-05, | |
| "loss": 0.3033, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.4058106841611997, | |
| "eval_loss": 0.07158821821212769, | |
| "eval_pearson_cosine": 0.7835729367892306, | |
| "eval_pearson_dot": 0.5987220055863389, | |
| "eval_pearson_euclidean": 0.7599913918565986, | |
| "eval_pearson_manhattan": 0.7589542766532986, | |
| "eval_runtime": 3.4365, | |
| "eval_samples_per_second": 436.486, | |
| "eval_spearman_cosine": 0.7961581681624044, | |
| "eval_spearman_dot": 0.5975812797826083, | |
| "eval_spearman_euclidean": 0.7727385023345612, | |
| "eval_spearman_manhattan": 0.7723228512152759, | |
| "eval_steps_per_second": 27.353, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.4104967197750704, | |
| "grad_norm": 2.3309245109558105, | |
| "learning_rate": 1.911843955014058e-05, | |
| "loss": 0.3238, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 1.415182755388941, | |
| "grad_norm": 2.3359642028808594, | |
| "learning_rate": 1.9115510777881914e-05, | |
| "loss": 0.2933, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.4198687910028116, | |
| "grad_norm": 2.363147497177124, | |
| "learning_rate": 1.9112582005623243e-05, | |
| "loss": 0.3204, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 1.4245548266166823, | |
| "grad_norm": 2.353273391723633, | |
| "learning_rate": 1.9109653233364576e-05, | |
| "loss": 0.295, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.429240862230553, | |
| "grad_norm": 2.418267011642456, | |
| "learning_rate": 1.9106724461105905e-05, | |
| "loss": 0.3397, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 1.4339268978444237, | |
| "grad_norm": 2.612607955932617, | |
| "learning_rate": 1.910379568884724e-05, | |
| "loss": 0.3001, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.4386129334582942, | |
| "grad_norm": 2.506488084793091, | |
| "learning_rate": 1.9100866916588568e-05, | |
| "loss": 0.3101, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 1.443298969072165, | |
| "grad_norm": 2.437333345413208, | |
| "learning_rate": 1.90979381443299e-05, | |
| "loss": 0.3017, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.4479850046860356, | |
| "grad_norm": 2.3391661643981934, | |
| "learning_rate": 1.909500937207123e-05, | |
| "loss": 0.2943, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 1.4526710402999063, | |
| "grad_norm": 2.410454511642456, | |
| "learning_rate": 1.909208059981256e-05, | |
| "loss": 0.3242, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.457357075913777, | |
| "grad_norm": 2.6086418628692627, | |
| "learning_rate": 1.9089151827553893e-05, | |
| "loss": 0.3442, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 1.4620431115276475, | |
| "grad_norm": 2.1769497394561768, | |
| "learning_rate": 1.9086223055295222e-05, | |
| "loss": 0.3114, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.4667291471415183, | |
| "grad_norm": 1.9582467079162598, | |
| "learning_rate": 1.908329428303655e-05, | |
| "loss": 0.3261, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 1.471415182755389, | |
| "grad_norm": 2.577211380004883, | |
| "learning_rate": 1.9080365510777884e-05, | |
| "loss": 0.31, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.4761012183692597, | |
| "grad_norm": 2.725773811340332, | |
| "learning_rate": 1.9077436738519214e-05, | |
| "loss": 0.2859, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 1.4807872539831304, | |
| "grad_norm": 2.779578685760498, | |
| "learning_rate": 1.9074507966260543e-05, | |
| "loss": 0.3304, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.4854732895970009, | |
| "grad_norm": 1.9493879079818726, | |
| "learning_rate": 1.9071579194001876e-05, | |
| "loss": 0.2992, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 1.4901593252108716, | |
| "grad_norm": 3.1011288166046143, | |
| "learning_rate": 1.9068650421743206e-05, | |
| "loss": 0.3406, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.4948453608247423, | |
| "grad_norm": 2.466146230697632, | |
| "learning_rate": 1.906572164948454e-05, | |
| "loss": 0.357, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 1.499531396438613, | |
| "grad_norm": 2.0382285118103027, | |
| "learning_rate": 1.9062792877225868e-05, | |
| "loss": 0.2968, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.5042174320524837, | |
| "grad_norm": 2.331540584564209, | |
| "learning_rate": 1.9059864104967198e-05, | |
| "loss": 0.3125, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 1.5089034676663542, | |
| "grad_norm": 2.590193271636963, | |
| "learning_rate": 1.905693533270853e-05, | |
| "loss": 0.2762, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.513589503280225, | |
| "grad_norm": 2.878938913345337, | |
| "learning_rate": 1.905400656044986e-05, | |
| "loss": 0.2868, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 1.5182755388940956, | |
| "grad_norm": 2.4432168006896973, | |
| "learning_rate": 1.9051077788191193e-05, | |
| "loss": 0.3021, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.522961574507966, | |
| "grad_norm": 2.849559783935547, | |
| "learning_rate": 1.9048149015932523e-05, | |
| "loss": 0.3247, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.522961574507966, | |
| "eval_loss": 0.07676094025373459, | |
| "eval_pearson_cosine": 0.7779113850636818, | |
| "eval_pearson_dot": 0.5637961042064976, | |
| "eval_pearson_euclidean": 0.7621398908224677, | |
| "eval_pearson_manhattan": 0.7612994290780648, | |
| "eval_runtime": 3.2908, | |
| "eval_samples_per_second": 455.817, | |
| "eval_spearman_cosine": 0.7910676412952927, | |
| "eval_spearman_dot": 0.562324725935799, | |
| "eval_spearman_euclidean": 0.7734903614875008, | |
| "eval_spearman_manhattan": 0.7731474604573987, | |
| "eval_steps_per_second": 28.565, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.527647610121837, | |
| "grad_norm": 2.372697591781616, | |
| "learning_rate": 1.9045220243673855e-05, | |
| "loss": 0.3423, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.5323336457357075, | |
| "grad_norm": 2.9068384170532227, | |
| "learning_rate": 1.9042291471415185e-05, | |
| "loss": 0.3133, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 1.5370196813495782, | |
| "grad_norm": 2.738391160964966, | |
| "learning_rate": 1.9039362699156518e-05, | |
| "loss": 0.3018, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.541705716963449, | |
| "grad_norm": 2.121777296066284, | |
| "learning_rate": 1.9036433926897847e-05, | |
| "loss": 0.3203, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 1.5463917525773194, | |
| "grad_norm": 2.1887810230255127, | |
| "learning_rate": 1.9033505154639177e-05, | |
| "loss": 0.3639, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.5510777881911904, | |
| "grad_norm": 2.4203438758850098, | |
| "learning_rate": 1.9030576382380506e-05, | |
| "loss": 0.3116, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 1.5557638238050608, | |
| "grad_norm": 2.00972318649292, | |
| "learning_rate": 1.902764761012184e-05, | |
| "loss": 0.3139, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.5604498594189316, | |
| "grad_norm": 2.093622922897339, | |
| "learning_rate": 1.902471883786317e-05, | |
| "loss": 0.3248, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 1.5651358950328023, | |
| "grad_norm": 2.5851595401763916, | |
| "learning_rate": 1.9021790065604498e-05, | |
| "loss": 0.3386, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.569821930646673, | |
| "grad_norm": 2.208651304244995, | |
| "learning_rate": 1.901886129334583e-05, | |
| "loss": 0.3138, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 1.5745079662605437, | |
| "grad_norm": 2.5663788318634033, | |
| "learning_rate": 1.901593252108716e-05, | |
| "loss": 0.3165, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.5791940018744142, | |
| "grad_norm": 2.443042039871216, | |
| "learning_rate": 1.9013003748828493e-05, | |
| "loss": 0.3178, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 1.5838800374882849, | |
| "grad_norm": 2.385019540786743, | |
| "learning_rate": 1.9010074976569823e-05, | |
| "loss": 0.2779, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.5885660731021556, | |
| "grad_norm": 2.5440433025360107, | |
| "learning_rate": 1.9007146204311156e-05, | |
| "loss": 0.3182, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 1.5932521087160263, | |
| "grad_norm": 2.1016016006469727, | |
| "learning_rate": 1.9004217432052485e-05, | |
| "loss": 0.3224, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.597938144329897, | |
| "grad_norm": 1.8776601552963257, | |
| "learning_rate": 1.9001288659793815e-05, | |
| "loss": 0.2926, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 1.6026241799437675, | |
| "grad_norm": 1.7855247259140015, | |
| "learning_rate": 1.8998359887535148e-05, | |
| "loss": 0.3087, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.6073102155576382, | |
| "grad_norm": 2.239816188812256, | |
| "learning_rate": 1.8995431115276477e-05, | |
| "loss": 0.3254, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 1.611996251171509, | |
| "grad_norm": 2.309274673461914, | |
| "learning_rate": 1.899250234301781e-05, | |
| "loss": 0.2741, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.6166822867853796, | |
| "grad_norm": 2.672464370727539, | |
| "learning_rate": 1.898957357075914e-05, | |
| "loss": 0.3137, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 1.6213683223992503, | |
| "grad_norm": 2.3761508464813232, | |
| "learning_rate": 1.8986644798500473e-05, | |
| "loss": 0.322, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.6260543580131208, | |
| "grad_norm": 2.497772216796875, | |
| "learning_rate": 1.8983716026241802e-05, | |
| "loss": 0.2832, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 1.6307403936269915, | |
| "grad_norm": 2.2633016109466553, | |
| "learning_rate": 1.898078725398313e-05, | |
| "loss": 0.2659, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.6354264292408622, | |
| "grad_norm": 2.2554588317871094, | |
| "learning_rate": 1.897785848172446e-05, | |
| "loss": 0.3039, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 1.640112464854733, | |
| "grad_norm": 2.0419113636016846, | |
| "learning_rate": 1.8974929709465794e-05, | |
| "loss": 0.26, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.640112464854733, | |
| "eval_loss": 0.06864778697490692, | |
| "eval_pearson_cosine": 0.7792468184875077, | |
| "eval_pearson_dot": 0.6003709070725307, | |
| "eval_pearson_euclidean": 0.762291216964627, | |
| "eval_pearson_manhattan": 0.7615198846806042, | |
| "eval_runtime": 3.2918, | |
| "eval_samples_per_second": 455.681, | |
| "eval_spearman_cosine": 0.7902077133733839, | |
| "eval_spearman_dot": 0.5998471986095437, | |
| "eval_spearman_euclidean": 0.7733862684662854, | |
| "eval_spearman_manhattan": 0.7732713082323596, | |
| "eval_steps_per_second": 28.556, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.6447985004686037, | |
| "grad_norm": 2.1808488368988037, | |
| "learning_rate": 1.8972000937207123e-05, | |
| "loss": 0.2977, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 1.6494845360824741, | |
| "grad_norm": 2.5023481845855713, | |
| "learning_rate": 1.8969072164948453e-05, | |
| "loss": 0.3275, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.6541705716963448, | |
| "grad_norm": 1.9074852466583252, | |
| "learning_rate": 1.8966143392689786e-05, | |
| "loss": 0.3084, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 1.6588566073102156, | |
| "grad_norm": 2.4666028022766113, | |
| "learning_rate": 1.8963214620431115e-05, | |
| "loss": 0.3202, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.6635426429240863, | |
| "grad_norm": 2.2712929248809814, | |
| "learning_rate": 1.8960285848172448e-05, | |
| "loss": 0.3023, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 1.668228678537957, | |
| "grad_norm": 2.274965286254883, | |
| "learning_rate": 1.8957357075913778e-05, | |
| "loss": 0.3134, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.6729147141518275, | |
| "grad_norm": 2.2933590412139893, | |
| "learning_rate": 1.895442830365511e-05, | |
| "loss": 0.2935, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 1.6776007497656982, | |
| "grad_norm": 2.492459774017334, | |
| "learning_rate": 1.895149953139644e-05, | |
| "loss": 0.2891, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.6822867853795689, | |
| "grad_norm": 2.4401142597198486, | |
| "learning_rate": 1.894857075913777e-05, | |
| "loss": 0.329, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 1.6869728209934396, | |
| "grad_norm": 2.0816047191619873, | |
| "learning_rate": 1.8945641986879102e-05, | |
| "loss": 0.3042, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.6916588566073103, | |
| "grad_norm": 2.2334861755371094, | |
| "learning_rate": 1.8942713214620432e-05, | |
| "loss": 0.2909, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 1.6963448922211808, | |
| "grad_norm": 2.3555166721343994, | |
| "learning_rate": 1.8939784442361765e-05, | |
| "loss": 0.3284, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.7010309278350515, | |
| "grad_norm": 1.751710295677185, | |
| "learning_rate": 1.8936855670103094e-05, | |
| "loss": 0.2848, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 1.7057169634489222, | |
| "grad_norm": 1.920424461364746, | |
| "learning_rate": 1.8933926897844427e-05, | |
| "loss": 0.279, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.710402999062793, | |
| "grad_norm": 2.840911865234375, | |
| "learning_rate": 1.8930998125585757e-05, | |
| "loss": 0.3116, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 1.7150890346766636, | |
| "grad_norm": 2.429884672164917, | |
| "learning_rate": 1.8928069353327086e-05, | |
| "loss": 0.264, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.7197750702905341, | |
| "grad_norm": 2.2484772205352783, | |
| "learning_rate": 1.892514058106842e-05, | |
| "loss": 0.3201, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 1.7244611059044048, | |
| "grad_norm": 2.342460870742798, | |
| "learning_rate": 1.892221180880975e-05, | |
| "loss": 0.3068, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 1.7291471415182755, | |
| "grad_norm": 2.332886219024658, | |
| "learning_rate": 1.8919283036551078e-05, | |
| "loss": 0.3238, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 1.7338331771321462, | |
| "grad_norm": 1.7878814935684204, | |
| "learning_rate": 1.8916354264292408e-05, | |
| "loss": 0.2866, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.738519212746017, | |
| "grad_norm": 2.455904722213745, | |
| "learning_rate": 1.891342549203374e-05, | |
| "loss": 0.2887, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 1.7432052483598874, | |
| "grad_norm": 2.273310899734497, | |
| "learning_rate": 1.891049671977507e-05, | |
| "loss": 0.3377, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 1.7478912839737581, | |
| "grad_norm": 2.77280592918396, | |
| "learning_rate": 1.8907567947516403e-05, | |
| "loss": 0.2928, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 1.7525773195876289, | |
| "grad_norm": 2.2917518615722656, | |
| "learning_rate": 1.8904639175257732e-05, | |
| "loss": 0.3154, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 1.7572633552014996, | |
| "grad_norm": 3.0059680938720703, | |
| "learning_rate": 1.8901710402999065e-05, | |
| "loss": 0.3216, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.7572633552014996, | |
| "eval_loss": 0.07067140191793442, | |
| "eval_pearson_cosine": 0.7850573661327331, | |
| "eval_pearson_dot": 0.6098455249231733, | |
| "eval_pearson_euclidean": 0.7677167356750871, | |
| "eval_pearson_manhattan": 0.7668202443827852, | |
| "eval_runtime": 3.5184, | |
| "eval_samples_per_second": 426.329, | |
| "eval_spearman_cosine": 0.7949681180333369, | |
| "eval_spearman_dot": 0.6135646781276878, | |
| "eval_spearman_euclidean": 0.7791366533328872, | |
| "eval_spearman_manhattan": 0.7786680488005274, | |
| "eval_steps_per_second": 26.717, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.7619493908153703, | |
| "grad_norm": 2.2580602169036865, | |
| "learning_rate": 1.8898781630740395e-05, | |
| "loss": 0.2838, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 1.7666354264292408, | |
| "grad_norm": 2.6639926433563232, | |
| "learning_rate": 1.8895852858481728e-05, | |
| "loss": 0.3388, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 1.7713214620431117, | |
| "grad_norm": 2.4491138458251953, | |
| "learning_rate": 1.8892924086223057e-05, | |
| "loss": 0.2905, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.7760074976569822, | |
| "grad_norm": 2.4412620067596436, | |
| "learning_rate": 1.8889995313964387e-05, | |
| "loss": 0.2982, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 1.780693533270853, | |
| "grad_norm": 2.356416702270508, | |
| "learning_rate": 1.888706654170572e-05, | |
| "loss": 0.3302, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.7853795688847236, | |
| "grad_norm": 2.3429980278015137, | |
| "learning_rate": 1.888413776944705e-05, | |
| "loss": 0.3069, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 1.790065604498594, | |
| "grad_norm": 2.228518486022949, | |
| "learning_rate": 1.8881208997188382e-05, | |
| "loss": 0.3016, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 1.794751640112465, | |
| "grad_norm": 2.8037281036376953, | |
| "learning_rate": 1.887828022492971e-05, | |
| "loss": 0.3324, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 1.7994376757263355, | |
| "grad_norm": 1.893836498260498, | |
| "learning_rate": 1.887535145267104e-05, | |
| "loss": 0.2592, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 1.8041237113402062, | |
| "grad_norm": 2.6253368854522705, | |
| "learning_rate": 1.8872422680412374e-05, | |
| "loss": 0.295, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 1.808809746954077, | |
| "grad_norm": 2.2410082817077637, | |
| "learning_rate": 1.8869493908153703e-05, | |
| "loss": 0.2495, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 1.8134957825679474, | |
| "grad_norm": 2.3530426025390625, | |
| "learning_rate": 1.8866565135895033e-05, | |
| "loss": 0.3004, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 2.178670644760132, | |
| "learning_rate": 1.8863636363636366e-05, | |
| "loss": 0.3386, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.8228678537956888, | |
| "grad_norm": 2.0660688877105713, | |
| "learning_rate": 1.8860707591377695e-05, | |
| "loss": 0.3117, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 1.8275538894095595, | |
| "grad_norm": 2.5521416664123535, | |
| "learning_rate": 1.8857778819119025e-05, | |
| "loss": 0.3011, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.8322399250234302, | |
| "grad_norm": 2.3752758502960205, | |
| "learning_rate": 1.8854850046860358e-05, | |
| "loss": 0.3203, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 1.8369259606373007, | |
| "grad_norm": 2.0836801528930664, | |
| "learning_rate": 1.8851921274601687e-05, | |
| "loss": 0.2706, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 1.8416119962511717, | |
| "grad_norm": 2.767982006072998, | |
| "learning_rate": 1.884899250234302e-05, | |
| "loss": 0.3077, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 1.8462980318650422, | |
| "grad_norm": 2.7820098400115967, | |
| "learning_rate": 1.884606373008435e-05, | |
| "loss": 0.2991, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 1.8509840674789129, | |
| "grad_norm": 2.6888327598571777, | |
| "learning_rate": 1.8843134957825682e-05, | |
| "loss": 0.2862, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 1.8556701030927836, | |
| "grad_norm": 2.4160315990448, | |
| "learning_rate": 1.8840206185567012e-05, | |
| "loss": 0.2725, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 1.860356138706654, | |
| "grad_norm": 2.778198719024658, | |
| "learning_rate": 1.8837277413308345e-05, | |
| "loss": 0.3509, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 1.865042174320525, | |
| "grad_norm": 1.8606864213943481, | |
| "learning_rate": 1.8834348641049674e-05, | |
| "loss": 0.2971, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 1.8697282099343955, | |
| "grad_norm": 1.98170804977417, | |
| "learning_rate": 1.8831419868791004e-05, | |
| "loss": 0.3131, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 1.8744142455482662, | |
| "grad_norm": 1.828123927116394, | |
| "learning_rate": 1.8828491096532337e-05, | |
| "loss": 0.3166, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.8744142455482662, | |
| "eval_loss": 0.07189416140317917, | |
| "eval_pearson_cosine": 0.7798801652938181, | |
| "eval_pearson_dot": 0.5736722449381091, | |
| "eval_pearson_euclidean": 0.7563095863468678, | |
| "eval_pearson_manhattan": 0.7550479478827015, | |
| "eval_runtime": 3.2996, | |
| "eval_samples_per_second": 454.607, | |
| "eval_spearman_cosine": 0.7910837741143455, | |
| "eval_spearman_dot": 0.5753555851409241, | |
| "eval_spearman_euclidean": 0.7700926689739851, | |
| "eval_spearman_manhattan": 0.7692694796240825, | |
| "eval_steps_per_second": 28.489, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.879100281162137, | |
| "grad_norm": 1.8813470602035522, | |
| "learning_rate": 1.8825562324273666e-05, | |
| "loss": 0.2484, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 1.8837863167760074, | |
| "grad_norm": 2.005892038345337, | |
| "learning_rate": 1.8822633552014996e-05, | |
| "loss": 0.3137, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 1.8884723523898783, | |
| "grad_norm": 2.067732572555542, | |
| "learning_rate": 1.881970477975633e-05, | |
| "loss": 0.3143, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 1.8931583880037488, | |
| "grad_norm": 2.33829665184021, | |
| "learning_rate": 1.8816776007497658e-05, | |
| "loss": 0.3252, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 1.8978444236176195, | |
| "grad_norm": 1.655529260635376, | |
| "learning_rate": 1.8813847235238987e-05, | |
| "loss": 0.2889, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 1.9025304592314902, | |
| "grad_norm": 2.849501609802246, | |
| "learning_rate": 1.881091846298032e-05, | |
| "loss": 0.323, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 1.9072164948453607, | |
| "grad_norm": 2.8345351219177246, | |
| "learning_rate": 1.880798969072165e-05, | |
| "loss": 0.316, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 1.9119025304592316, | |
| "grad_norm": 2.061378240585327, | |
| "learning_rate": 1.8805060918462983e-05, | |
| "loss": 0.3112, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 1.9165885660731021, | |
| "grad_norm": 1.883039116859436, | |
| "learning_rate": 1.8802132146204312e-05, | |
| "loss": 0.2625, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 1.9212746016869728, | |
| "grad_norm": 2.2609167098999023, | |
| "learning_rate": 1.8799203373945642e-05, | |
| "loss": 0.3043, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.9259606373008435, | |
| "grad_norm": 2.27116060256958, | |
| "learning_rate": 1.8796274601686975e-05, | |
| "loss": 0.3022, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 1.930646672914714, | |
| "grad_norm": 1.9829683303833008, | |
| "learning_rate": 1.8793345829428304e-05, | |
| "loss": 0.3147, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 1.935332708528585, | |
| "grad_norm": 2.073610305786133, | |
| "learning_rate": 1.8790417057169637e-05, | |
| "loss": 0.2779, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 1.9400187441424555, | |
| "grad_norm": 2.183823585510254, | |
| "learning_rate": 1.8787488284910967e-05, | |
| "loss": 0.3041, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 1.9447047797563262, | |
| "grad_norm": 2.373502016067505, | |
| "learning_rate": 1.87845595126523e-05, | |
| "loss": 0.3037, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 1.9493908153701969, | |
| "grad_norm": 3.190950393676758, | |
| "learning_rate": 1.878163074039363e-05, | |
| "loss": 0.2876, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 1.9540768509840674, | |
| "grad_norm": 2.2904293537139893, | |
| "learning_rate": 1.8778701968134962e-05, | |
| "loss": 0.2957, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 1.9587628865979383, | |
| "grad_norm": 2.5620875358581543, | |
| "learning_rate": 1.877577319587629e-05, | |
| "loss": 0.2865, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 1.9634489222118088, | |
| "grad_norm": 1.8882808685302734, | |
| "learning_rate": 1.877284442361762e-05, | |
| "loss": 0.302, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 1.9681349578256795, | |
| "grad_norm": 2.736853837966919, | |
| "learning_rate": 1.876991565135895e-05, | |
| "loss": 0.2783, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.9728209934395502, | |
| "grad_norm": 2.6126515865325928, | |
| "learning_rate": 1.8766986879100283e-05, | |
| "loss": 0.3337, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 1.9775070290534207, | |
| "grad_norm": 1.9608129262924194, | |
| "learning_rate": 1.8764058106841613e-05, | |
| "loss": 0.2817, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 1.9821930646672916, | |
| "grad_norm": 2.7429044246673584, | |
| "learning_rate": 1.8761129334582942e-05, | |
| "loss": 0.2827, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 1.986879100281162, | |
| "grad_norm": 2.322411298751831, | |
| "learning_rate": 1.8758200562324275e-05, | |
| "loss": 0.2967, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 1.9915651358950328, | |
| "grad_norm": 1.80522882938385, | |
| "learning_rate": 1.8755271790065605e-05, | |
| "loss": 0.315, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.9915651358950328, | |
| "eval_loss": 0.0709637850522995, | |
| "eval_pearson_cosine": 0.7818087092412611, | |
| "eval_pearson_dot": 0.5917647670204929, | |
| "eval_pearson_euclidean": 0.7671767812906438, | |
| "eval_pearson_manhattan": 0.7657443892799449, | |
| "eval_runtime": 3.3617, | |
| "eval_samples_per_second": 446.198, | |
| "eval_spearman_cosine": 0.7925261968709182, | |
| "eval_spearman_dot": 0.5930341342679433, | |
| "eval_spearman_euclidean": 0.7790026457642986, | |
| "eval_spearman_manhattan": 0.7779600937996545, | |
| "eval_steps_per_second": 27.962, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.9962511715089035, | |
| "grad_norm": 2.3154516220092773, | |
| "learning_rate": 1.8752343017806937e-05, | |
| "loss": 0.3019, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 2.000937207122774, | |
| "grad_norm": 1.769853949546814, | |
| "learning_rate": 1.8749414245548267e-05, | |
| "loss": 0.2639, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 2.005623242736645, | |
| "grad_norm": 2.7089197635650635, | |
| "learning_rate": 1.87464854732896e-05, | |
| "loss": 0.1872, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 2.0103092783505154, | |
| "grad_norm": 2.1212291717529297, | |
| "learning_rate": 1.874355670103093e-05, | |
| "loss": 0.189, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 2.014995313964386, | |
| "grad_norm": 1.8745218515396118, | |
| "learning_rate": 1.874062792877226e-05, | |
| "loss": 0.1749, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 2.019681349578257, | |
| "grad_norm": 2.0145514011383057, | |
| "learning_rate": 1.8737699156513592e-05, | |
| "loss": 0.1809, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 2.0243673851921273, | |
| "grad_norm": 2.370055913925171, | |
| "learning_rate": 1.873477038425492e-05, | |
| "loss": 0.2056, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 2.0290534208059983, | |
| "grad_norm": 1.7560367584228516, | |
| "learning_rate": 1.8731841611996254e-05, | |
| "loss": 0.1742, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 2.0337394564198688, | |
| "grad_norm": 2.693603277206421, | |
| "learning_rate": 1.8728912839737584e-05, | |
| "loss": 0.2017, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 2.0384254920337392, | |
| "grad_norm": 2.1456475257873535, | |
| "learning_rate": 1.8725984067478917e-05, | |
| "loss": 0.1967, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 2.04311152764761, | |
| "grad_norm": 1.8832343816757202, | |
| "learning_rate": 1.8723055295220246e-05, | |
| "loss": 0.182, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 2.0477975632614807, | |
| "grad_norm": 2.2446205615997314, | |
| "learning_rate": 1.8720126522961576e-05, | |
| "loss": 0.1895, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 2.0524835988753516, | |
| "grad_norm": 1.8662675619125366, | |
| "learning_rate": 1.871719775070291e-05, | |
| "loss": 0.2091, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 2.057169634489222, | |
| "grad_norm": 2.3427059650421143, | |
| "learning_rate": 1.8714268978444238e-05, | |
| "loss": 0.2046, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 2.0618556701030926, | |
| "grad_norm": 1.8921583890914917, | |
| "learning_rate": 1.8711340206185567e-05, | |
| "loss": 0.21, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 2.0665417057169635, | |
| "grad_norm": 2.3345112800598145, | |
| "learning_rate": 1.8708411433926897e-05, | |
| "loss": 0.211, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 2.071227741330834, | |
| "grad_norm": 2.8094024658203125, | |
| "learning_rate": 1.870548266166823e-05, | |
| "loss": 0.185, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 2.075913776944705, | |
| "grad_norm": 2.182375192642212, | |
| "learning_rate": 1.870255388940956e-05, | |
| "loss": 0.1743, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 2.0805998125585754, | |
| "grad_norm": 1.9309943914413452, | |
| "learning_rate": 1.8699625117150892e-05, | |
| "loss": 0.2002, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 2.085285848172446, | |
| "grad_norm": 2.410891532897949, | |
| "learning_rate": 1.869669634489222e-05, | |
| "loss": 0.2034, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 2.089971883786317, | |
| "grad_norm": 1.9030256271362305, | |
| "learning_rate": 1.8693767572633555e-05, | |
| "loss": 0.1985, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 2.0946579194001873, | |
| "grad_norm": 2.387491226196289, | |
| "learning_rate": 1.8690838800374884e-05, | |
| "loss": 0.1657, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 2.0993439550140582, | |
| "grad_norm": 2.448557138442993, | |
| "learning_rate": 1.8687910028116217e-05, | |
| "loss": 0.1889, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 2.1040299906279287, | |
| "grad_norm": 1.8442336320877075, | |
| "learning_rate": 1.8684981255857546e-05, | |
| "loss": 0.1638, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 2.108716026241799, | |
| "grad_norm": 2.192814588546753, | |
| "learning_rate": 1.8682052483598876e-05, | |
| "loss": 0.2117, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.108716026241799, | |
| "eval_loss": 0.0544867068529129, | |
| "eval_pearson_cosine": 0.7772149225219422, | |
| "eval_pearson_dot": 0.6058722441499071, | |
| "eval_pearson_euclidean": 0.7566861769282127, | |
| "eval_pearson_manhattan": 0.7550653692259459, | |
| "eval_runtime": 3.562, | |
| "eval_samples_per_second": 421.112, | |
| "eval_spearman_cosine": 0.7889882574404774, | |
| "eval_spearman_dot": 0.609585139530218, | |
| "eval_spearman_euclidean": 0.7711664957300913, | |
| "eval_spearman_manhattan": 0.7702480077346849, | |
| "eval_steps_per_second": 26.39, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.11340206185567, | |
| "grad_norm": 1.964051604270935, | |
| "learning_rate": 1.867912371134021e-05, | |
| "loss": 0.1727, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 2.1180880974695406, | |
| "grad_norm": 2.180359363555908, | |
| "learning_rate": 1.867619493908154e-05, | |
| "loss": 0.1929, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 2.1227741330834116, | |
| "grad_norm": 1.5920389890670776, | |
| "learning_rate": 1.867326616682287e-05, | |
| "loss": 0.215, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 2.127460168697282, | |
| "grad_norm": 2.0330374240875244, | |
| "learning_rate": 1.86703373945642e-05, | |
| "loss": 0.1782, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 2.1321462043111525, | |
| "grad_norm": 3.1396799087524414, | |
| "learning_rate": 1.866740862230553e-05, | |
| "loss": 0.1843, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 2.1368322399250235, | |
| "grad_norm": 2.507819652557373, | |
| "learning_rate": 1.8664479850046863e-05, | |
| "loss": 0.2272, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 2.141518275538894, | |
| "grad_norm": 1.9377321004867554, | |
| "learning_rate": 1.8661551077788193e-05, | |
| "loss": 0.2003, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 2.146204311152765, | |
| "grad_norm": 2.868741512298584, | |
| "learning_rate": 1.8658622305529522e-05, | |
| "loss": 0.2266, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 2.1508903467666354, | |
| "grad_norm": 1.8727847337722778, | |
| "learning_rate": 1.8655693533270855e-05, | |
| "loss": 0.1965, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 2.155576382380506, | |
| "grad_norm": 1.937414526939392, | |
| "learning_rate": 1.8652764761012184e-05, | |
| "loss": 0.1957, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 2.160262417994377, | |
| "grad_norm": 2.084174156188965, | |
| "learning_rate": 1.8649835988753514e-05, | |
| "loss": 0.1668, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 2.1649484536082473, | |
| "grad_norm": 1.846552848815918, | |
| "learning_rate": 1.8646907216494847e-05, | |
| "loss": 0.184, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 2.169634489222118, | |
| "grad_norm": 1.9087766408920288, | |
| "learning_rate": 1.8643978444236176e-05, | |
| "loss": 0.1961, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 2.1743205248359887, | |
| "grad_norm": 2.011108875274658, | |
| "learning_rate": 1.864104967197751e-05, | |
| "loss": 0.1848, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 2.179006560449859, | |
| "grad_norm": 2.8976380825042725, | |
| "learning_rate": 1.863812089971884e-05, | |
| "loss": 0.2082, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 2.18369259606373, | |
| "grad_norm": 2.1731081008911133, | |
| "learning_rate": 1.863519212746017e-05, | |
| "loss": 0.1852, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 2.1883786316776006, | |
| "grad_norm": 1.7672990560531616, | |
| "learning_rate": 1.86322633552015e-05, | |
| "loss": 0.2023, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 2.1930646672914715, | |
| "grad_norm": 1.6886131763458252, | |
| "learning_rate": 1.8629334582942834e-05, | |
| "loss": 0.2126, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 2.197750702905342, | |
| "grad_norm": 1.4913769960403442, | |
| "learning_rate": 1.8626405810684164e-05, | |
| "loss": 0.1836, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 2.2024367385192125, | |
| "grad_norm": 2.0632176399230957, | |
| "learning_rate": 1.8623477038425493e-05, | |
| "loss": 0.2048, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 2.2071227741330834, | |
| "grad_norm": 1.9084508419036865, | |
| "learning_rate": 1.8620548266166826e-05, | |
| "loss": 0.1836, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 2.211808809746954, | |
| "grad_norm": 1.9352916479110718, | |
| "learning_rate": 1.8617619493908155e-05, | |
| "loss": 0.1653, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 2.216494845360825, | |
| "grad_norm": 1.974461317062378, | |
| "learning_rate": 1.8614690721649485e-05, | |
| "loss": 0.2053, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 2.2211808809746953, | |
| "grad_norm": 2.1307456493377686, | |
| "learning_rate": 1.8611761949390818e-05, | |
| "loss": 0.1835, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 2.2258669165885663, | |
| "grad_norm": 1.413037896156311, | |
| "learning_rate": 1.8608833177132147e-05, | |
| "loss": 0.1725, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 2.2258669165885663, | |
| "eval_loss": 0.05439988151192665, | |
| "eval_pearson_cosine": 0.7779719017710534, | |
| "eval_pearson_dot": 0.6064920766481805, | |
| "eval_pearson_euclidean": 0.7604967084059453, | |
| "eval_pearson_manhattan": 0.7592805758538006, | |
| "eval_runtime": 3.2739, | |
| "eval_samples_per_second": 458.175, | |
| "eval_spearman_cosine": 0.786759628149343, | |
| "eval_spearman_dot": 0.6127818290418755, | |
| "eval_spearman_euclidean": 0.7721037448944629, | |
| "eval_spearman_manhattan": 0.7714373415381524, | |
| "eval_steps_per_second": 28.712, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 2.2305529522024368, | |
| "grad_norm": 2.033710241317749, | |
| "learning_rate": 1.8605904404873477e-05, | |
| "loss": 0.1769, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 2.2352389878163073, | |
| "grad_norm": 2.065793752670288, | |
| "learning_rate": 1.860297563261481e-05, | |
| "loss": 0.1896, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 2.239925023430178, | |
| "grad_norm": 2.0393176078796387, | |
| "learning_rate": 1.860004686035614e-05, | |
| "loss": 0.1944, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 2.2446110590440487, | |
| "grad_norm": 1.967092514038086, | |
| "learning_rate": 1.859711808809747e-05, | |
| "loss": 0.218, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 2.2492970946579196, | |
| "grad_norm": 2.614811420440674, | |
| "learning_rate": 1.85941893158388e-05, | |
| "loss": 0.2158, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 2.25398313027179, | |
| "grad_norm": 1.9915626049041748, | |
| "learning_rate": 1.859126054358013e-05, | |
| "loss": 0.199, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 2.2586691658856606, | |
| "grad_norm": 2.308933973312378, | |
| "learning_rate": 1.8588331771321464e-05, | |
| "loss": 0.1746, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 2.2633552014995315, | |
| "grad_norm": 2.2984583377838135, | |
| "learning_rate": 1.8585402999062793e-05, | |
| "loss": 0.1905, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 2.268041237113402, | |
| "grad_norm": 2.0273101329803467, | |
| "learning_rate": 1.8582474226804126e-05, | |
| "loss": 0.1655, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 2.2727272727272725, | |
| "grad_norm": 1.745671272277832, | |
| "learning_rate": 1.8579545454545456e-05, | |
| "loss": 0.1745, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 2.2774133083411434, | |
| "grad_norm": 1.9413400888442993, | |
| "learning_rate": 1.857661668228679e-05, | |
| "loss": 0.1899, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 2.282099343955014, | |
| "grad_norm": 2.113260269165039, | |
| "learning_rate": 1.8573687910028118e-05, | |
| "loss": 0.1826, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 2.286785379568885, | |
| "grad_norm": 1.9468377828598022, | |
| "learning_rate": 1.857075913776945e-05, | |
| "loss": 0.1596, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 2.2914714151827553, | |
| "grad_norm": 2.1250803470611572, | |
| "learning_rate": 1.856783036551078e-05, | |
| "loss": 0.1969, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 2.296157450796626, | |
| "grad_norm": 1.5394418239593506, | |
| "learning_rate": 1.856490159325211e-05, | |
| "loss": 0.1731, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 2.3008434864104967, | |
| "grad_norm": 2.2445003986358643, | |
| "learning_rate": 1.8561972820993443e-05, | |
| "loss": 0.1817, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 2.3055295220243672, | |
| "grad_norm": 2.201733350753784, | |
| "learning_rate": 1.8559044048734773e-05, | |
| "loss": 0.175, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 2.310215557638238, | |
| "grad_norm": 2.077092409133911, | |
| "learning_rate": 1.8556115276476102e-05, | |
| "loss": 0.1755, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 2.3149015932521086, | |
| "grad_norm": 2.0630977153778076, | |
| "learning_rate": 1.855318650421743e-05, | |
| "loss": 0.1967, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 2.319587628865979, | |
| "grad_norm": 2.737861156463623, | |
| "learning_rate": 1.8550257731958764e-05, | |
| "loss": 0.2175, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 2.32427366447985, | |
| "grad_norm": 2.0301871299743652, | |
| "learning_rate": 1.8547328959700094e-05, | |
| "loss": 0.1938, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 2.3289597000937206, | |
| "grad_norm": 1.6436868906021118, | |
| "learning_rate": 1.8544400187441427e-05, | |
| "loss": 0.2215, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 2.3336457357075915, | |
| "grad_norm": 2.0046422481536865, | |
| "learning_rate": 1.8541471415182756e-05, | |
| "loss": 0.189, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 2.338331771321462, | |
| "grad_norm": 2.598639488220215, | |
| "learning_rate": 1.8538542642924086e-05, | |
| "loss": 0.2075, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 2.3430178069353325, | |
| "grad_norm": 1.8523390293121338, | |
| "learning_rate": 1.853561387066542e-05, | |
| "loss": 0.1985, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.3430178069353325, | |
| "eval_loss": 0.05401456356048584, | |
| "eval_pearson_cosine": 0.781767118943602, | |
| "eval_pearson_dot": 0.6017329721420595, | |
| "eval_pearson_euclidean": 0.7626066877478763, | |
| "eval_pearson_manhattan": 0.7620550203728129, | |
| "eval_runtime": 3.2064, | |
| "eval_samples_per_second": 467.808, | |
| "eval_spearman_cosine": 0.7915644905642973, | |
| "eval_spearman_dot": 0.6077979618715047, | |
| "eval_spearman_euclidean": 0.773425356513352, | |
| "eval_spearman_manhattan": 0.7732819218811388, | |
| "eval_steps_per_second": 29.316, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.3477038425492034, | |
| "grad_norm": 2.0773608684539795, | |
| "learning_rate": 1.8532685098406748e-05, | |
| "loss": 0.1749, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 2.352389878163074, | |
| "grad_norm": 1.6482021808624268, | |
| "learning_rate": 1.852975632614808e-05, | |
| "loss": 0.1719, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 2.357075913776945, | |
| "grad_norm": 2.1038990020751953, | |
| "learning_rate": 1.852682755388941e-05, | |
| "loss": 0.2014, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 2.3617619493908153, | |
| "grad_norm": 2.396871566772461, | |
| "learning_rate": 1.8523898781630743e-05, | |
| "loss": 0.1855, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 2.3664479850046862, | |
| "grad_norm": 2.3098223209381104, | |
| "learning_rate": 1.8520970009372073e-05, | |
| "loss": 0.1819, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 2.3711340206185567, | |
| "grad_norm": 1.9691858291625977, | |
| "learning_rate": 1.8518041237113406e-05, | |
| "loss": 0.1786, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 2.375820056232427, | |
| "grad_norm": 2.8055005073547363, | |
| "learning_rate": 1.8515112464854735e-05, | |
| "loss": 0.1864, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 2.380506091846298, | |
| "grad_norm": 2.2753820419311523, | |
| "learning_rate": 1.8512183692596065e-05, | |
| "loss": 0.1918, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 2.3851921274601686, | |
| "grad_norm": 1.632279634475708, | |
| "learning_rate": 1.8509254920337398e-05, | |
| "loss": 0.1767, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 2.3898781630740396, | |
| "grad_norm": 2.177457571029663, | |
| "learning_rate": 1.8506326148078727e-05, | |
| "loss": 0.1866, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 2.39456419868791, | |
| "grad_norm": 1.8641494512557983, | |
| "learning_rate": 1.8503397375820057e-05, | |
| "loss": 0.1833, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 2.3992502343017805, | |
| "grad_norm": 2.0116262435913086, | |
| "learning_rate": 1.8500468603561386e-05, | |
| "loss": 0.1975, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 2.4039362699156515, | |
| "grad_norm": 1.130846381187439, | |
| "learning_rate": 1.849753983130272e-05, | |
| "loss": 0.1992, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 2.408622305529522, | |
| "grad_norm": 2.743990421295166, | |
| "learning_rate": 1.849461105904405e-05, | |
| "loss": 0.1996, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 2.413308341143393, | |
| "grad_norm": 2.4368762969970703, | |
| "learning_rate": 1.849168228678538e-05, | |
| "loss": 0.1722, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 2.4179943767572634, | |
| "grad_norm": 2.659679889678955, | |
| "learning_rate": 1.848875351452671e-05, | |
| "loss": 0.1834, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 2.422680412371134, | |
| "grad_norm": 1.9601231813430786, | |
| "learning_rate": 1.8485824742268044e-05, | |
| "loss": 0.1943, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 2.427366447985005, | |
| "grad_norm": 1.521337628364563, | |
| "learning_rate": 1.8482895970009373e-05, | |
| "loss": 0.2009, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 2.4320524835988753, | |
| "grad_norm": 1.4766628742218018, | |
| "learning_rate": 1.8479967197750703e-05, | |
| "loss": 0.1848, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 2.436738519212746, | |
| "grad_norm": 2.3502743244171143, | |
| "learning_rate": 1.8477038425492036e-05, | |
| "loss": 0.165, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 2.4414245548266167, | |
| "grad_norm": 1.5100184679031372, | |
| "learning_rate": 1.8474109653233365e-05, | |
| "loss": 0.1983, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 2.446110590440487, | |
| "grad_norm": 2.336106061935425, | |
| "learning_rate": 1.8471180880974698e-05, | |
| "loss": 0.1871, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 2.450796626054358, | |
| "grad_norm": 2.1695876121520996, | |
| "learning_rate": 1.8468252108716028e-05, | |
| "loss": 0.1996, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 2.4554826616682286, | |
| "grad_norm": 2.047522783279419, | |
| "learning_rate": 1.846532333645736e-05, | |
| "loss": 0.2088, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 2.4601686972820995, | |
| "grad_norm": 1.9469248056411743, | |
| "learning_rate": 1.846239456419869e-05, | |
| "loss": 0.1871, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 2.4601686972820995, | |
| "eval_loss": 0.05267513543367386, | |
| "eval_pearson_cosine": 0.7830377509621513, | |
| "eval_pearson_dot": 0.5843463467596735, | |
| "eval_pearson_euclidean": 0.7586592990276699, | |
| "eval_pearson_manhattan": 0.7576312827364555, | |
| "eval_runtime": 3.3095, | |
| "eval_samples_per_second": 453.24, | |
| "eval_spearman_cosine": 0.7898144287463136, | |
| "eval_spearman_dot": 0.5893690662486756, | |
| "eval_spearman_euclidean": 0.7723820863953171, | |
| "eval_spearman_manhattan": 0.7718409919252442, | |
| "eval_steps_per_second": 28.403, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 2.46485473289597, | |
| "grad_norm": 2.1970055103302, | |
| "learning_rate": 1.845946579194002e-05, | |
| "loss": 0.2059, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 2.4695407685098405, | |
| "grad_norm": 1.7079726457595825, | |
| "learning_rate": 1.8456537019681352e-05, | |
| "loss": 0.1677, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 2.4742268041237114, | |
| "grad_norm": 2.1129045486450195, | |
| "learning_rate": 1.8453608247422682e-05, | |
| "loss": 0.1774, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 2.478912839737582, | |
| "grad_norm": 2.0570807456970215, | |
| "learning_rate": 1.845067947516401e-05, | |
| "loss": 0.1863, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 2.483598875351453, | |
| "grad_norm": 2.095052719116211, | |
| "learning_rate": 1.8447750702905344e-05, | |
| "loss": 0.1675, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 2.4882849109653233, | |
| "grad_norm": 2.5408618450164795, | |
| "learning_rate": 1.8444821930646674e-05, | |
| "loss": 0.2033, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 2.492970946579194, | |
| "grad_norm": 2.0885772705078125, | |
| "learning_rate": 1.8441893158388003e-05, | |
| "loss": 0.178, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 2.4976569821930648, | |
| "grad_norm": 2.907564640045166, | |
| "learning_rate": 1.8438964386129336e-05, | |
| "loss": 0.2121, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 2.5023430178069352, | |
| "grad_norm": 1.3953640460968018, | |
| "learning_rate": 1.8436035613870666e-05, | |
| "loss": 0.1838, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 2.5070290534208057, | |
| "grad_norm": 1.825296401977539, | |
| "learning_rate": 1.8433106841612e-05, | |
| "loss": 0.1688, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 2.5117150890346767, | |
| "grad_norm": 2.5563881397247314, | |
| "learning_rate": 1.8430178069353328e-05, | |
| "loss": 0.1798, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 2.5164011246485476, | |
| "grad_norm": 2.355055570602417, | |
| "learning_rate": 1.842724929709466e-05, | |
| "loss": 0.1858, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 2.521087160262418, | |
| "grad_norm": 2.4546165466308594, | |
| "learning_rate": 1.842432052483599e-05, | |
| "loss": 0.2149, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 2.5257731958762886, | |
| "grad_norm": 2.0621817111968994, | |
| "learning_rate": 1.842139175257732e-05, | |
| "loss": 0.2215, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 2.530459231490159, | |
| "grad_norm": 1.7873655557632446, | |
| "learning_rate": 1.8418462980318653e-05, | |
| "loss": 0.1872, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 2.53514526710403, | |
| "grad_norm": 1.8408267498016357, | |
| "learning_rate": 1.8415534208059982e-05, | |
| "loss": 0.181, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 2.539831302717901, | |
| "grad_norm": 2.56502628326416, | |
| "learning_rate": 1.8412605435801315e-05, | |
| "loss": 0.1946, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 2.5445173383317714, | |
| "grad_norm": 1.9909722805023193, | |
| "learning_rate": 1.8409676663542645e-05, | |
| "loss": 0.1912, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 2.549203373945642, | |
| "grad_norm": 2.530014753341675, | |
| "learning_rate": 1.8406747891283974e-05, | |
| "loss": 0.1898, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 2.5538894095595124, | |
| "grad_norm": 2.452223300933838, | |
| "learning_rate": 1.8403819119025307e-05, | |
| "loss": 0.2114, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 2.5585754451733833, | |
| "grad_norm": 2.0805137157440186, | |
| "learning_rate": 1.8400890346766637e-05, | |
| "loss": 0.2237, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 2.5632614807872542, | |
| "grad_norm": 1.8469469547271729, | |
| "learning_rate": 1.8397961574507966e-05, | |
| "loss": 0.1895, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 2.5679475164011247, | |
| "grad_norm": 2.0643043518066406, | |
| "learning_rate": 1.83950328022493e-05, | |
| "loss": 0.1943, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 2.572633552014995, | |
| "grad_norm": 2.1368064880371094, | |
| "learning_rate": 1.839210402999063e-05, | |
| "loss": 0.1851, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 2.5773195876288657, | |
| "grad_norm": 1.7404872179031372, | |
| "learning_rate": 1.8389175257731958e-05, | |
| "loss": 0.17, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.5773195876288657, | |
| "eval_loss": 0.052087847143411636, | |
| "eval_pearson_cosine": 0.7876996404495316, | |
| "eval_pearson_dot": 0.6239791899062368, | |
| "eval_pearson_euclidean": 0.7632876525798107, | |
| "eval_pearson_manhattan": 0.7621358820486535, | |
| "eval_runtime": 3.2114, | |
| "eval_samples_per_second": 467.081, | |
| "eval_spearman_cosine": 0.7959069730064968, | |
| "eval_spearman_dot": 0.6246328836004849, | |
| "eval_spearman_euclidean": 0.775279722994119, | |
| "eval_spearman_manhattan": 0.7746119015682993, | |
| "eval_steps_per_second": 29.27, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.5820056232427366, | |
| "grad_norm": 1.7516288757324219, | |
| "learning_rate": 1.838624648547329e-05, | |
| "loss": 0.2103, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 2.5866916588566076, | |
| "grad_norm": 2.0095157623291016, | |
| "learning_rate": 1.838331771321462e-05, | |
| "loss": 0.2149, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 2.591377694470478, | |
| "grad_norm": 2.3408539295196533, | |
| "learning_rate": 1.8380388940955953e-05, | |
| "loss": 0.1737, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 2.5960637300843485, | |
| "grad_norm": 2.1319546699523926, | |
| "learning_rate": 1.8377460168697283e-05, | |
| "loss": 0.1963, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 2.600749765698219, | |
| "grad_norm": 1.6769325733184814, | |
| "learning_rate": 1.8374531396438616e-05, | |
| "loss": 0.1953, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 2.60543580131209, | |
| "grad_norm": 2.503340244293213, | |
| "learning_rate": 1.8371602624179945e-05, | |
| "loss": 0.1872, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 2.610121836925961, | |
| "grad_norm": 1.9402929544448853, | |
| "learning_rate": 1.8368673851921278e-05, | |
| "loss": 0.2101, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 2.6148078725398314, | |
| "grad_norm": 1.980652928352356, | |
| "learning_rate": 1.8365745079662608e-05, | |
| "loss": 0.1847, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 2.619493908153702, | |
| "grad_norm": 2.6117632389068604, | |
| "learning_rate": 1.8362816307403937e-05, | |
| "loss": 0.1934, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 2.624179943767573, | |
| "grad_norm": 1.8052171468734741, | |
| "learning_rate": 1.835988753514527e-05, | |
| "loss": 0.1827, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 2.6288659793814433, | |
| "grad_norm": 2.384441375732422, | |
| "learning_rate": 1.83569587628866e-05, | |
| "loss": 0.1765, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 2.633552014995314, | |
| "grad_norm": 1.8288124799728394, | |
| "learning_rate": 1.8354029990627932e-05, | |
| "loss": 0.189, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 2.6382380506091847, | |
| "grad_norm": 1.8904608488082886, | |
| "learning_rate": 1.8351101218369262e-05, | |
| "loss": 0.1858, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 2.642924086223055, | |
| "grad_norm": 2.4750778675079346, | |
| "learning_rate": 1.834817244611059e-05, | |
| "loss": 0.2139, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 2.647610121836926, | |
| "grad_norm": 1.642340898513794, | |
| "learning_rate": 1.834524367385192e-05, | |
| "loss": 0.2069, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 2.6522961574507966, | |
| "grad_norm": 2.656212091445923, | |
| "learning_rate": 1.8342314901593254e-05, | |
| "loss": 0.1975, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 2.6569821930646675, | |
| "grad_norm": 1.8371754884719849, | |
| "learning_rate": 1.8339386129334583e-05, | |
| "loss": 0.2044, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 2.661668228678538, | |
| "grad_norm": 1.4980151653289795, | |
| "learning_rate": 1.8336457357075916e-05, | |
| "loss": 0.1641, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 2.6663542642924085, | |
| "grad_norm": 2.300854206085205, | |
| "learning_rate": 1.8333528584817246e-05, | |
| "loss": 0.22, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 2.6710402999062794, | |
| "grad_norm": 1.6950479745864868, | |
| "learning_rate": 1.8330599812558575e-05, | |
| "loss": 0.194, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 2.67572633552015, | |
| "grad_norm": 2.4695844650268555, | |
| "learning_rate": 1.8327671040299908e-05, | |
| "loss": 0.2099, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 2.680412371134021, | |
| "grad_norm": 1.6376041173934937, | |
| "learning_rate": 1.8324742268041237e-05, | |
| "loss": 0.1844, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 2.6850984067478914, | |
| "grad_norm": 2.3736367225646973, | |
| "learning_rate": 1.832181349578257e-05, | |
| "loss": 0.1886, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 2.689784442361762, | |
| "grad_norm": 1.9710888862609863, | |
| "learning_rate": 1.83188847235239e-05, | |
| "loss": 0.1881, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 2.6944704779756328, | |
| "grad_norm": 2.38665771484375, | |
| "learning_rate": 1.8315955951265233e-05, | |
| "loss": 0.174, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 2.6944704779756328, | |
| "eval_loss": 0.052811067551374435, | |
| "eval_pearson_cosine": 0.7875700428566148, | |
| "eval_pearson_dot": 0.6195988091254279, | |
| "eval_pearson_euclidean": 0.7602836109774316, | |
| "eval_pearson_manhattan": 0.7593741208437699, | |
| "eval_runtime": 3.3995, | |
| "eval_samples_per_second": 441.245, | |
| "eval_spearman_cosine": 0.7949301691809272, | |
| "eval_spearman_dot": 0.6234473775126543, | |
| "eval_spearman_euclidean": 0.771585595419442, | |
| "eval_spearman_manhattan": 0.7712513119620406, | |
| "eval_steps_per_second": 27.651, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 2.6991565135895033, | |
| "grad_norm": 1.9556349515914917, | |
| "learning_rate": 1.8313027179006562e-05, | |
| "loss": 0.1984, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 2.703842549203374, | |
| "grad_norm": 1.8807729482650757, | |
| "learning_rate": 1.8310098406747895e-05, | |
| "loss": 0.2057, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 2.7085285848172447, | |
| "grad_norm": 1.74807870388031, | |
| "learning_rate": 1.8307169634489225e-05, | |
| "loss": 0.2158, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 2.713214620431115, | |
| "grad_norm": 2.3334672451019287, | |
| "learning_rate": 1.8304240862230554e-05, | |
| "loss": 0.1676, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 2.717900656044986, | |
| "grad_norm": 2.0100314617156982, | |
| "learning_rate": 1.8301312089971887e-05, | |
| "loss": 0.1785, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 2.7225866916588566, | |
| "grad_norm": 1.9850749969482422, | |
| "learning_rate": 1.8298383317713217e-05, | |
| "loss": 0.191, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 2.7272727272727275, | |
| "grad_norm": 2.1841342449188232, | |
| "learning_rate": 1.8295454545454546e-05, | |
| "loss": 0.1814, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 2.731958762886598, | |
| "grad_norm": 2.6331992149353027, | |
| "learning_rate": 1.8292525773195876e-05, | |
| "loss": 0.2218, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 2.7366447985004685, | |
| "grad_norm": 2.161170482635498, | |
| "learning_rate": 1.828959700093721e-05, | |
| "loss": 0.1834, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 2.7413308341143394, | |
| "grad_norm": 2.3174726963043213, | |
| "learning_rate": 1.8286668228678538e-05, | |
| "loss": 0.1863, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 2.74601686972821, | |
| "grad_norm": 1.9110337495803833, | |
| "learning_rate": 1.828373945641987e-05, | |
| "loss": 0.179, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 2.750702905342081, | |
| "grad_norm": 2.59794545173645, | |
| "learning_rate": 1.82808106841612e-05, | |
| "loss": 0.1891, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 2.7553889409559513, | |
| "grad_norm": 2.1763062477111816, | |
| "learning_rate": 1.8277881911902533e-05, | |
| "loss": 0.2094, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 2.760074976569822, | |
| "grad_norm": 2.3349921703338623, | |
| "learning_rate": 1.8274953139643863e-05, | |
| "loss": 0.1848, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 2.7647610121836927, | |
| "grad_norm": 2.5409889221191406, | |
| "learning_rate": 1.8272024367385192e-05, | |
| "loss": 0.2392, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 2.7694470477975632, | |
| "grad_norm": 2.0501179695129395, | |
| "learning_rate": 1.8269095595126525e-05, | |
| "loss": 0.2116, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 2.774133083411434, | |
| "grad_norm": 2.486163854598999, | |
| "learning_rate": 1.8266166822867855e-05, | |
| "loss": 0.1899, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 2.7788191190253047, | |
| "grad_norm": 2.43495774269104, | |
| "learning_rate": 1.8263238050609187e-05, | |
| "loss": 0.212, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 2.783505154639175, | |
| "grad_norm": 1.9861983060836792, | |
| "learning_rate": 1.8260309278350517e-05, | |
| "loss": 0.2135, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 2.788191190253046, | |
| "grad_norm": 2.46321964263916, | |
| "learning_rate": 1.825738050609185e-05, | |
| "loss": 0.2033, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 2.7928772258669166, | |
| "grad_norm": 2.2375664710998535, | |
| "learning_rate": 1.825445173383318e-05, | |
| "loss": 0.2483, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 2.7975632614807875, | |
| "grad_norm": 2.2775328159332275, | |
| "learning_rate": 1.825152296157451e-05, | |
| "loss": 0.1921, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 2.802249297094658, | |
| "grad_norm": 2.1675665378570557, | |
| "learning_rate": 1.8248594189315842e-05, | |
| "loss": 0.1666, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 2.8069353327085285, | |
| "grad_norm": 1.969119668006897, | |
| "learning_rate": 1.824566541705717e-05, | |
| "loss": 0.1997, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 2.8116213683223994, | |
| "grad_norm": 3.1404953002929688, | |
| "learning_rate": 1.82427366447985e-05, | |
| "loss": 0.1896, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.8116213683223994, | |
| "eval_loss": 0.0506160594522953, | |
| "eval_pearson_cosine": 0.7847906228898438, | |
| "eval_pearson_dot": 0.6051671435295134, | |
| "eval_pearson_euclidean": 0.7605907930754796, | |
| "eval_pearson_manhattan": 0.7595079560474858, | |
| "eval_runtime": 3.1727, | |
| "eval_samples_per_second": 472.783, | |
| "eval_spearman_cosine": 0.7891227097506602, | |
| "eval_spearman_dot": 0.6083246161085648, | |
| "eval_spearman_euclidean": 0.7718421311199012, | |
| "eval_spearman_manhattan": 0.7711946620901435, | |
| "eval_steps_per_second": 29.628, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.81630740393627, | |
| "grad_norm": 2.5197956562042236, | |
| "learning_rate": 1.8239807872539834e-05, | |
| "loss": 0.19, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 2.820993439550141, | |
| "grad_norm": 2.342001438140869, | |
| "learning_rate": 1.8236879100281163e-05, | |
| "loss": 0.1902, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 2.8256794751640113, | |
| "grad_norm": 2.081195592880249, | |
| "learning_rate": 1.8233950328022493e-05, | |
| "loss": 0.1861, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 2.830365510777882, | |
| "grad_norm": 2.2222466468811035, | |
| "learning_rate": 1.8231021555763825e-05, | |
| "loss": 0.1956, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 2.8350515463917527, | |
| "grad_norm": 2.531655788421631, | |
| "learning_rate": 1.8228092783505155e-05, | |
| "loss": 0.2001, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 2.839737582005623, | |
| "grad_norm": 1.8432163000106812, | |
| "learning_rate": 1.8225164011246488e-05, | |
| "loss": 0.1917, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 2.844423617619494, | |
| "grad_norm": 1.8071563243865967, | |
| "learning_rate": 1.8222235238987817e-05, | |
| "loss": 0.2292, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 2.8491096532333646, | |
| "grad_norm": 2.596858501434326, | |
| "learning_rate": 1.821930646672915e-05, | |
| "loss": 0.2025, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 2.853795688847235, | |
| "grad_norm": 2.516810655593872, | |
| "learning_rate": 1.821637769447048e-05, | |
| "loss": 0.2008, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 2.858481724461106, | |
| "grad_norm": 2.685643196105957, | |
| "learning_rate": 1.821344892221181e-05, | |
| "loss": 0.2181, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 2.8631677600749765, | |
| "grad_norm": 2.3013880252838135, | |
| "learning_rate": 1.8210520149953142e-05, | |
| "loss": 0.2089, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 2.8678537956888475, | |
| "grad_norm": 2.1551058292388916, | |
| "learning_rate": 1.820759137769447e-05, | |
| "loss": 0.2096, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 2.872539831302718, | |
| "grad_norm": 2.2167651653289795, | |
| "learning_rate": 1.8204662605435805e-05, | |
| "loss": 0.1928, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 2.8772258669165884, | |
| "grad_norm": 2.3424344062805176, | |
| "learning_rate": 1.8201733833177134e-05, | |
| "loss": 0.1843, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 2.8819119025304594, | |
| "grad_norm": 2.8250808715820312, | |
| "learning_rate": 1.8198805060918464e-05, | |
| "loss": 0.2221, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 2.88659793814433, | |
| "grad_norm": 2.3614280223846436, | |
| "learning_rate": 1.8195876288659796e-05, | |
| "loss": 0.1862, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 2.891283973758201, | |
| "grad_norm": 1.918118953704834, | |
| "learning_rate": 1.8192947516401126e-05, | |
| "loss": 0.1867, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 2.8959700093720713, | |
| "grad_norm": 1.8377277851104736, | |
| "learning_rate": 1.8190018744142455e-05, | |
| "loss": 0.1826, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 2.9006560449859418, | |
| "grad_norm": 2.527493715286255, | |
| "learning_rate": 1.8187089971883788e-05, | |
| "loss": 0.1986, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 2.9053420805998127, | |
| "grad_norm": 2.125455379486084, | |
| "learning_rate": 1.8184161199625118e-05, | |
| "loss": 0.1936, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 2.910028116213683, | |
| "grad_norm": 2.2447023391723633, | |
| "learning_rate": 1.8181232427366447e-05, | |
| "loss": 0.1887, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 2.914714151827554, | |
| "grad_norm": 2.3427064418792725, | |
| "learning_rate": 1.817830365510778e-05, | |
| "loss": 0.197, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 2.9194001874414246, | |
| "grad_norm": 2.480989933013916, | |
| "learning_rate": 1.817537488284911e-05, | |
| "loss": 0.2145, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 2.924086223055295, | |
| "grad_norm": 2.2610814571380615, | |
| "learning_rate": 1.8172446110590443e-05, | |
| "loss": 0.1854, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 2.928772258669166, | |
| "grad_norm": 2.220090866088867, | |
| "learning_rate": 1.8169517338331772e-05, | |
| "loss": 0.1897, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 2.928772258669166, | |
| "eval_loss": 0.05493560805916786, | |
| "eval_pearson_cosine": 0.7819311073703084, | |
| "eval_pearson_dot": 0.5956676516410369, | |
| "eval_pearson_euclidean": 0.7532898404318757, | |
| "eval_pearson_manhattan": 0.7521288710513865, | |
| "eval_runtime": 3.273, | |
| "eval_samples_per_second": 458.299, | |
| "eval_spearman_cosine": 0.7902389528249426, | |
| "eval_spearman_dot": 0.5980515739599185, | |
| "eval_spearman_euclidean": 0.7666646754982684, | |
| "eval_spearman_manhattan": 0.7664381416282928, | |
| "eval_steps_per_second": 28.72, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 2.9334582942830365, | |
| "grad_norm": 2.517014741897583, | |
| "learning_rate": 1.8166588566073105e-05, | |
| "loss": 0.2132, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 2.9381443298969074, | |
| "grad_norm": 1.9690264463424683, | |
| "learning_rate": 1.8163659793814434e-05, | |
| "loss": 0.1756, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 2.942830365510778, | |
| "grad_norm": 1.812945008277893, | |
| "learning_rate": 1.8160731021555767e-05, | |
| "loss": 0.1848, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 2.9475164011246484, | |
| "grad_norm": 2.003582000732422, | |
| "learning_rate": 1.8157802249297097e-05, | |
| "loss": 0.2071, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 2.9522024367385193, | |
| "grad_norm": 3.1531078815460205, | |
| "learning_rate": 1.8154873477038426e-05, | |
| "loss": 0.2146, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 2.95688847235239, | |
| "grad_norm": 2.5869102478027344, | |
| "learning_rate": 1.815194470477976e-05, | |
| "loss": 0.1854, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 2.9615745079662608, | |
| "grad_norm": 2.2422142028808594, | |
| "learning_rate": 1.814901593252109e-05, | |
| "loss": 0.179, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 2.9662605435801312, | |
| "grad_norm": 2.346750497817993, | |
| "learning_rate": 1.814608716026242e-05, | |
| "loss": 0.1937, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 2.9709465791940017, | |
| "grad_norm": 1.6241106986999512, | |
| "learning_rate": 1.814315838800375e-05, | |
| "loss": 0.1945, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 2.9756326148078727, | |
| "grad_norm": 2.8439619541168213, | |
| "learning_rate": 1.814022961574508e-05, | |
| "loss": 0.2265, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 2.980318650421743, | |
| "grad_norm": 1.69356107711792, | |
| "learning_rate": 1.813730084348641e-05, | |
| "loss": 0.1869, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 2.985004686035614, | |
| "grad_norm": 2.3386266231536865, | |
| "learning_rate": 1.8134372071227743e-05, | |
| "loss": 0.1752, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 2.9896907216494846, | |
| "grad_norm": 1.873734712600708, | |
| "learning_rate": 1.8131443298969072e-05, | |
| "loss": 0.1972, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 2.994376757263355, | |
| "grad_norm": 2.3732969760894775, | |
| "learning_rate": 1.8128514526710402e-05, | |
| "loss": 0.2229, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 2.999062792877226, | |
| "grad_norm": 2.2805166244506836, | |
| "learning_rate": 1.8125585754451735e-05, | |
| "loss": 0.2022, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 3.0037488284910965, | |
| "grad_norm": 1.8053455352783203, | |
| "learning_rate": 1.8122656982193064e-05, | |
| "loss": 0.132, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 3.0084348641049674, | |
| "grad_norm": 2.322193145751953, | |
| "learning_rate": 1.8119728209934397e-05, | |
| "loss": 0.1341, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 3.013120899718838, | |
| "grad_norm": 2.2462315559387207, | |
| "learning_rate": 1.8116799437675727e-05, | |
| "loss": 0.1294, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 3.0178069353327084, | |
| "grad_norm": 1.5314648151397705, | |
| "learning_rate": 1.811387066541706e-05, | |
| "loss": 0.0974, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 3.0224929709465793, | |
| "grad_norm": 1.4587002992630005, | |
| "learning_rate": 1.811094189315839e-05, | |
| "loss": 0.1019, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 3.02717900656045, | |
| "grad_norm": 0.887192964553833, | |
| "learning_rate": 1.8108013120899722e-05, | |
| "loss": 0.0945, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 3.0318650421743207, | |
| "grad_norm": 1.8087823390960693, | |
| "learning_rate": 1.810508434864105e-05, | |
| "loss": 0.1103, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 3.036551077788191, | |
| "grad_norm": 2.4967281818389893, | |
| "learning_rate": 1.8102155576382384e-05, | |
| "loss": 0.1261, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 3.0412371134020617, | |
| "grad_norm": 1.4955415725708008, | |
| "learning_rate": 1.8099226804123714e-05, | |
| "loss": 0.1255, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 3.0459231490159326, | |
| "grad_norm": 1.0758466720581055, | |
| "learning_rate": 1.8096298031865043e-05, | |
| "loss": 0.105, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 3.0459231490159326, | |
| "eval_loss": 0.04503355920314789, | |
| "eval_pearson_cosine": 0.7887213700331159, | |
| "eval_pearson_dot": 0.638523586851079, | |
| "eval_pearson_euclidean": 0.7527277991826509, | |
| "eval_pearson_manhattan": 0.7516127412435569, | |
| "eval_runtime": 3.2262, | |
| "eval_samples_per_second": 464.937, | |
| "eval_spearman_cosine": 0.7931480987271526, | |
| "eval_spearman_dot": 0.6449590450062397, | |
| "eval_spearman_euclidean": 0.7674534939783395, | |
| "eval_spearman_manhattan": 0.7668609796988295, | |
| "eval_steps_per_second": 29.136, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 3.050609184629803, | |
| "grad_norm": 2.3318862915039062, | |
| "learning_rate": 1.8093369259606376e-05, | |
| "loss": 0.1364, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 3.055295220243674, | |
| "grad_norm": 1.8145966529846191, | |
| "learning_rate": 1.8090440487347706e-05, | |
| "loss": 0.1236, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 3.0599812558575445, | |
| "grad_norm": 1.584845781326294, | |
| "learning_rate": 1.8087511715089035e-05, | |
| "loss": 0.128, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 3.064667291471415, | |
| "grad_norm": 1.2164894342422485, | |
| "learning_rate": 1.8084582942830368e-05, | |
| "loss": 0.1233, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 3.069353327085286, | |
| "grad_norm": 1.761679768562317, | |
| "learning_rate": 1.8081654170571698e-05, | |
| "loss": 0.1107, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 3.0740393626991565, | |
| "grad_norm": 1.9818497896194458, | |
| "learning_rate": 1.8078725398313027e-05, | |
| "loss": 0.1373, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 3.0787253983130274, | |
| "grad_norm": 2.124472141265869, | |
| "learning_rate": 1.807579662605436e-05, | |
| "loss": 0.1409, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 3.083411433926898, | |
| "grad_norm": 1.6288917064666748, | |
| "learning_rate": 1.807286785379569e-05, | |
| "loss": 0.1022, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 3.0880974695407684, | |
| "grad_norm": 1.5154041051864624, | |
| "learning_rate": 1.806993908153702e-05, | |
| "loss": 0.1167, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 3.0927835051546393, | |
| "grad_norm": 1.736624836921692, | |
| "learning_rate": 1.8067010309278352e-05, | |
| "loss": 0.0932, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 3.0974695407685098, | |
| "grad_norm": 2.2292051315307617, | |
| "learning_rate": 1.806408153701968e-05, | |
| "loss": 0.1458, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 3.1021555763823807, | |
| "grad_norm": 1.8836901187896729, | |
| "learning_rate": 1.8061152764761014e-05, | |
| "loss": 0.1145, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 3.106841611996251, | |
| "grad_norm": 2.0602309703826904, | |
| "learning_rate": 1.8058223992502344e-05, | |
| "loss": 0.1091, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 3.1115276476101217, | |
| "grad_norm": 1.7996163368225098, | |
| "learning_rate": 1.8055295220243677e-05, | |
| "loss": 0.1155, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 3.1162136832239926, | |
| "grad_norm": 1.695610761642456, | |
| "learning_rate": 1.8052366447985006e-05, | |
| "loss": 0.1268, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 3.120899718837863, | |
| "grad_norm": 1.6327197551727295, | |
| "learning_rate": 1.804943767572634e-05, | |
| "loss": 0.1157, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 3.125585754451734, | |
| "grad_norm": 2.457167387008667, | |
| "learning_rate": 1.804650890346767e-05, | |
| "loss": 0.1348, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 3.1302717900656045, | |
| "grad_norm": 1.1226837635040283, | |
| "learning_rate": 1.8043580131208998e-05, | |
| "loss": 0.1112, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 3.134957825679475, | |
| "grad_norm": 1.904137134552002, | |
| "learning_rate": 1.804065135895033e-05, | |
| "loss": 0.1288, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 3.139643861293346, | |
| "grad_norm": 1.8711892366409302, | |
| "learning_rate": 1.803772258669166e-05, | |
| "loss": 0.1341, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 3.1443298969072164, | |
| "grad_norm": 1.8008291721343994, | |
| "learning_rate": 1.803479381443299e-05, | |
| "loss": 0.1198, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 3.1490159325210874, | |
| "grad_norm": 1.8362482786178589, | |
| "learning_rate": 1.8031865042174323e-05, | |
| "loss": 0.1479, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 3.153701968134958, | |
| "grad_norm": 1.6150809526443481, | |
| "learning_rate": 1.8028936269915652e-05, | |
| "loss": 0.1138, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 3.1583880037488283, | |
| "grad_norm": 2.2239766120910645, | |
| "learning_rate": 1.8026007497656982e-05, | |
| "loss": 0.1309, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 3.1630740393626993, | |
| "grad_norm": 1.669233798980713, | |
| "learning_rate": 1.8023078725398315e-05, | |
| "loss": 0.1055, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 3.1630740393626993, | |
| "eval_loss": 0.04599650204181671, | |
| "eval_pearson_cosine": 0.7875010466521104, | |
| "eval_pearson_dot": 0.6256249285068609, | |
| "eval_pearson_euclidean": 0.752459338375715, | |
| "eval_pearson_manhattan": 0.7515050600000066, | |
| "eval_runtime": 3.3759, | |
| "eval_samples_per_second": 444.332, | |
| "eval_spearman_cosine": 0.792665874705252, | |
| "eval_spearman_dot": 0.6331527343675145, | |
| "eval_spearman_euclidean": 0.7657151594858318, | |
| "eval_spearman_manhattan": 0.7651890994431135, | |
| "eval_steps_per_second": 27.845, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 3.1677600749765698, | |
| "grad_norm": 2.1133861541748047, | |
| "learning_rate": 1.8020149953139644e-05, | |
| "loss": 0.1128, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 3.1724461105904407, | |
| "grad_norm": 1.5470513105392456, | |
| "learning_rate": 1.8017221180880977e-05, | |
| "loss": 0.1246, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 3.177132146204311, | |
| "grad_norm": 1.9087189435958862, | |
| "learning_rate": 1.8014292408622307e-05, | |
| "loss": 0.1137, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 3.1818181818181817, | |
| "grad_norm": 1.9185723066329956, | |
| "learning_rate": 1.8011363636363636e-05, | |
| "loss": 0.1196, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 3.1865042174320526, | |
| "grad_norm": 1.7757681608200073, | |
| "learning_rate": 1.800843486410497e-05, | |
| "loss": 0.0978, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 3.191190253045923, | |
| "grad_norm": 1.6667931079864502, | |
| "learning_rate": 1.80055060918463e-05, | |
| "loss": 0.113, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 3.195876288659794, | |
| "grad_norm": 2.101773500442505, | |
| "learning_rate": 1.800257731958763e-05, | |
| "loss": 0.1034, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 3.2005623242736645, | |
| "grad_norm": 1.6078181266784668, | |
| "learning_rate": 1.799964854732896e-05, | |
| "loss": 0.1276, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 3.205248359887535, | |
| "grad_norm": 1.3552807569503784, | |
| "learning_rate": 1.7996719775070294e-05, | |
| "loss": 0.1184, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 3.209934395501406, | |
| "grad_norm": 1.5063031911849976, | |
| "learning_rate": 1.7993791002811623e-05, | |
| "loss": 0.1367, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 3.2146204311152764, | |
| "grad_norm": 1.9200366735458374, | |
| "learning_rate": 1.7990862230552956e-05, | |
| "loss": 0.1222, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 3.2193064667291473, | |
| "grad_norm": 1.4902859926223755, | |
| "learning_rate": 1.7987933458294286e-05, | |
| "loss": 0.1414, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 3.223992502343018, | |
| "grad_norm": 1.7316768169403076, | |
| "learning_rate": 1.7985004686035615e-05, | |
| "loss": 0.1451, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 3.2286785379568883, | |
| "grad_norm": 1.8446965217590332, | |
| "learning_rate": 1.7982075913776945e-05, | |
| "loss": 0.1227, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 3.2333645735707592, | |
| "grad_norm": 1.5571191310882568, | |
| "learning_rate": 1.7979147141518278e-05, | |
| "loss": 0.1193, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 3.2380506091846297, | |
| "grad_norm": 2.3433127403259277, | |
| "learning_rate": 1.7976218369259607e-05, | |
| "loss": 0.1175, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 3.2427366447985007, | |
| "grad_norm": 1.8916155099868774, | |
| "learning_rate": 1.7973289597000937e-05, | |
| "loss": 0.1352, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 3.247422680412371, | |
| "grad_norm": 1.3708059787750244, | |
| "learning_rate": 1.797036082474227e-05, | |
| "loss": 0.1174, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 3.2521087160262416, | |
| "grad_norm": 1.5119550228118896, | |
| "learning_rate": 1.79674320524836e-05, | |
| "loss": 0.1331, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 3.2567947516401126, | |
| "grad_norm": 1.8884799480438232, | |
| "learning_rate": 1.7964503280224932e-05, | |
| "loss": 0.125, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 3.261480787253983, | |
| "grad_norm": 1.6222647428512573, | |
| "learning_rate": 1.796157450796626e-05, | |
| "loss": 0.1191, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 3.266166822867854, | |
| "grad_norm": 1.2746548652648926, | |
| "learning_rate": 1.7958645735707594e-05, | |
| "loss": 0.1256, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 3.2708528584817245, | |
| "grad_norm": 1.883239984512329, | |
| "learning_rate": 1.7955716963448924e-05, | |
| "loss": 0.1283, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 3.275538894095595, | |
| "grad_norm": 2.0994622707366943, | |
| "learning_rate": 1.7952788191190253e-05, | |
| "loss": 0.1316, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 3.280224929709466, | |
| "grad_norm": 1.844297170639038, | |
| "learning_rate": 1.7949859418931586e-05, | |
| "loss": 0.1145, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 3.280224929709466, | |
| "eval_loss": 0.045286137610673904, | |
| "eval_pearson_cosine": 0.7924534055849843, | |
| "eval_pearson_dot": 0.6315508963322145, | |
| "eval_pearson_euclidean": 0.7559423098506386, | |
| "eval_pearson_manhattan": 0.7547566595358202, | |
| "eval_runtime": 3.1537, | |
| "eval_samples_per_second": 475.63, | |
| "eval_spearman_cosine": 0.7977173793324885, | |
| "eval_spearman_dot": 0.6408281685024612, | |
| "eval_spearman_euclidean": 0.7678066060450297, | |
| "eval_spearman_manhattan": 0.7671464837523794, | |
| "eval_steps_per_second": 29.806, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 3.2849109653233364, | |
| "grad_norm": 1.3316779136657715, | |
| "learning_rate": 1.7946930646672916e-05, | |
| "loss": 0.1139, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 3.2895970009372073, | |
| "grad_norm": 1.3782374858856201, | |
| "learning_rate": 1.794400187441425e-05, | |
| "loss": 0.1238, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 3.294283036551078, | |
| "grad_norm": 1.4582282304763794, | |
| "learning_rate": 1.7941073102155578e-05, | |
| "loss": 0.1279, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 3.2989690721649483, | |
| "grad_norm": 2.3581652641296387, | |
| "learning_rate": 1.793814432989691e-05, | |
| "loss": 0.118, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 3.303655107778819, | |
| "grad_norm": 2.1047656536102295, | |
| "learning_rate": 1.793521555763824e-05, | |
| "loss": 0.1086, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 3.3083411433926897, | |
| "grad_norm": 1.748353123664856, | |
| "learning_rate": 1.793228678537957e-05, | |
| "loss": 0.1357, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 3.3130271790065606, | |
| "grad_norm": 1.905609130859375, | |
| "learning_rate": 1.79293580131209e-05, | |
| "loss": 0.1239, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 3.317713214620431, | |
| "grad_norm": 1.5777533054351807, | |
| "learning_rate": 1.7926429240862232e-05, | |
| "loss": 0.1179, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 3.3223992502343016, | |
| "grad_norm": 1.8429454565048218, | |
| "learning_rate": 1.7923500468603562e-05, | |
| "loss": 0.1103, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 3.3270852858481725, | |
| "grad_norm": 1.601769208908081, | |
| "learning_rate": 1.792057169634489e-05, | |
| "loss": 0.1282, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 3.331771321462043, | |
| "grad_norm": 1.948339581489563, | |
| "learning_rate": 1.7917642924086224e-05, | |
| "loss": 0.1314, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 3.336457357075914, | |
| "grad_norm": 2.035210609436035, | |
| "learning_rate": 1.7914714151827554e-05, | |
| "loss": 0.1232, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 3.3411433926897844, | |
| "grad_norm": 1.3871088027954102, | |
| "learning_rate": 1.7911785379568887e-05, | |
| "loss": 0.1225, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 3.345829428303655, | |
| "grad_norm": 1.9721599817276, | |
| "learning_rate": 1.7908856607310216e-05, | |
| "loss": 0.1364, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 3.350515463917526, | |
| "grad_norm": 2.2410221099853516, | |
| "learning_rate": 1.790592783505155e-05, | |
| "loss": 0.1277, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 3.3552014995313963, | |
| "grad_norm": 1.8568767309188843, | |
| "learning_rate": 1.790299906279288e-05, | |
| "loss": 0.1114, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 3.3598875351452673, | |
| "grad_norm": 2.367185115814209, | |
| "learning_rate": 1.790007029053421e-05, | |
| "loss": 0.1334, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 3.3645735707591378, | |
| "grad_norm": 1.6528865098953247, | |
| "learning_rate": 1.789714151827554e-05, | |
| "loss": 0.1223, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 3.3692596063730083, | |
| "grad_norm": 1.9136816263198853, | |
| "learning_rate": 1.789421274601687e-05, | |
| "loss": 0.1066, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 3.373945641986879, | |
| "grad_norm": 2.2870426177978516, | |
| "learning_rate": 1.7891283973758203e-05, | |
| "loss": 0.1058, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 3.3786316776007497, | |
| "grad_norm": 1.5177735090255737, | |
| "learning_rate": 1.7888355201499533e-05, | |
| "loss": 0.1179, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 3.3833177132146206, | |
| "grad_norm": 2.215899705886841, | |
| "learning_rate": 1.7885426429240866e-05, | |
| "loss": 0.1321, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 3.388003748828491, | |
| "grad_norm": 1.7164250612258911, | |
| "learning_rate": 1.7882497656982195e-05, | |
| "loss": 0.1111, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 3.3926897844423616, | |
| "grad_norm": 2.3435146808624268, | |
| "learning_rate": 1.7879568884723525e-05, | |
| "loss": 0.1513, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 3.3973758200562325, | |
| "grad_norm": 1.6385475397109985, | |
| "learning_rate": 1.7876640112464858e-05, | |
| "loss": 0.1252, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 3.3973758200562325, | |
| "eval_loss": 0.047000955790281296, | |
| "eval_pearson_cosine": 0.7888804465638501, | |
| "eval_pearson_dot": 0.6256929197655836, | |
| "eval_pearson_euclidean": 0.7570957556748965, | |
| "eval_pearson_manhattan": 0.7561323835241964, | |
| "eval_runtime": 3.4083, | |
| "eval_samples_per_second": 440.106, | |
| "eval_spearman_cosine": 0.7947042550328448, | |
| "eval_spearman_dot": 0.6282838912238756, | |
| "eval_spearman_euclidean": 0.7692784528325559, | |
| "eval_spearman_manhattan": 0.7683456146298958, | |
| "eval_steps_per_second": 27.58, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 3.402061855670103, | |
| "grad_norm": 1.5348442792892456, | |
| "learning_rate": 1.7873711340206187e-05, | |
| "loss": 0.1304, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 3.406747891283974, | |
| "grad_norm": 1.854861855506897, | |
| "learning_rate": 1.7870782567947517e-05, | |
| "loss": 0.1296, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 3.4114339268978444, | |
| "grad_norm": 1.5103728771209717, | |
| "learning_rate": 1.7867853795688846e-05, | |
| "loss": 0.1182, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 3.416119962511715, | |
| "grad_norm": 1.4604400396347046, | |
| "learning_rate": 1.786492502343018e-05, | |
| "loss": 0.1155, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 3.420805998125586, | |
| "grad_norm": 1.9087272882461548, | |
| "learning_rate": 1.786199625117151e-05, | |
| "loss": 0.1265, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 3.4254920337394563, | |
| "grad_norm": 2.3286941051483154, | |
| "learning_rate": 1.785906747891284e-05, | |
| "loss": 0.0926, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 3.4301780693533273, | |
| "grad_norm": 1.8551883697509766, | |
| "learning_rate": 1.785613870665417e-05, | |
| "loss": 0.122, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 3.4348641049671977, | |
| "grad_norm": 1.7142447233200073, | |
| "learning_rate": 1.7853209934395504e-05, | |
| "loss": 0.138, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 3.4395501405810682, | |
| "grad_norm": 1.5548756122589111, | |
| "learning_rate": 1.7850281162136833e-05, | |
| "loss": 0.1123, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 3.444236176194939, | |
| "grad_norm": 1.3273271322250366, | |
| "learning_rate": 1.7847352389878166e-05, | |
| "loss": 0.129, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 3.4489222118088096, | |
| "grad_norm": 2.410999298095703, | |
| "learning_rate": 1.7844423617619496e-05, | |
| "loss": 0.1329, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 3.4536082474226806, | |
| "grad_norm": 2.0272533893585205, | |
| "learning_rate": 1.784149484536083e-05, | |
| "loss": 0.1251, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 3.458294283036551, | |
| "grad_norm": 2.4070749282836914, | |
| "learning_rate": 1.7838566073102158e-05, | |
| "loss": 0.1384, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 3.4629803186504216, | |
| "grad_norm": 2.0983331203460693, | |
| "learning_rate": 1.7835637300843487e-05, | |
| "loss": 0.1268, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 3.4676663542642925, | |
| "grad_norm": 1.5650460720062256, | |
| "learning_rate": 1.783270852858482e-05, | |
| "loss": 0.1218, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 3.472352389878163, | |
| "grad_norm": 1.4505574703216553, | |
| "learning_rate": 1.782977975632615e-05, | |
| "loss": 0.1278, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 3.477038425492034, | |
| "grad_norm": 2.4104747772216797, | |
| "learning_rate": 1.782685098406748e-05, | |
| "loss": 0.1506, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 3.4817244611059044, | |
| "grad_norm": 2.2749085426330566, | |
| "learning_rate": 1.7823922211808812e-05, | |
| "loss": 0.1382, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 3.486410496719775, | |
| "grad_norm": 1.619983434677124, | |
| "learning_rate": 1.7820993439550142e-05, | |
| "loss": 0.1145, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 3.491096532333646, | |
| "grad_norm": 1.797600269317627, | |
| "learning_rate": 1.781806466729147e-05, | |
| "loss": 0.1381, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 3.4957825679475163, | |
| "grad_norm": 1.2865104675292969, | |
| "learning_rate": 1.7815135895032804e-05, | |
| "loss": 0.1088, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 3.5004686035613872, | |
| "grad_norm": 2.2955312728881836, | |
| "learning_rate": 1.7812207122774134e-05, | |
| "loss": 0.1389, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 3.5051546391752577, | |
| "grad_norm": 1.884758472442627, | |
| "learning_rate": 1.7809278350515463e-05, | |
| "loss": 0.1399, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 3.509840674789128, | |
| "grad_norm": 2.0076119899749756, | |
| "learning_rate": 1.7806349578256796e-05, | |
| "loss": 0.1028, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 3.514526710402999, | |
| "grad_norm": 1.5606528520584106, | |
| "learning_rate": 1.7803420805998125e-05, | |
| "loss": 0.1058, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 3.514526710402999, | |
| "eval_loss": 0.044608332216739655, | |
| "eval_pearson_cosine": 0.7913143880230251, | |
| "eval_pearson_dot": 0.6221286850907433, | |
| "eval_pearson_euclidean": 0.7578325537672654, | |
| "eval_pearson_manhattan": 0.7571791901529252, | |
| "eval_runtime": 3.3012, | |
| "eval_samples_per_second": 454.374, | |
| "eval_spearman_cosine": 0.7958246101441605, | |
| "eval_spearman_dot": 0.6338421877202784, | |
| "eval_spearman_euclidean": 0.7715109007041803, | |
| "eval_spearman_manhattan": 0.7714185275105551, | |
| "eval_steps_per_second": 28.474, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 3.5192127460168696, | |
| "grad_norm": 1.9054310321807861, | |
| "learning_rate": 1.780049203373946e-05, | |
| "loss": 0.1149, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 3.5238987816307406, | |
| "grad_norm": 2.2473042011260986, | |
| "learning_rate": 1.7797563261480788e-05, | |
| "loss": 0.1241, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 3.528584817244611, | |
| "grad_norm": 2.9301676750183105, | |
| "learning_rate": 1.779463448922212e-05, | |
| "loss": 0.1155, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 3.5332708528584815, | |
| "grad_norm": 1.8250157833099365, | |
| "learning_rate": 1.779170571696345e-05, | |
| "loss": 0.118, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 3.5379568884723525, | |
| "grad_norm": 2.101794958114624, | |
| "learning_rate": 1.7788776944704783e-05, | |
| "loss": 0.1338, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 3.542642924086223, | |
| "grad_norm": 1.9058297872543335, | |
| "learning_rate": 1.7785848172446113e-05, | |
| "loss": 0.1172, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 3.547328959700094, | |
| "grad_norm": 2.1106228828430176, | |
| "learning_rate": 1.7782919400187446e-05, | |
| "loss": 0.119, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 3.5520149953139644, | |
| "grad_norm": 1.572771430015564, | |
| "learning_rate": 1.7779990627928775e-05, | |
| "loss": 0.1303, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 3.556701030927835, | |
| "grad_norm": 2.670691967010498, | |
| "learning_rate": 1.7777061855670105e-05, | |
| "loss": 0.1367, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 3.561387066541706, | |
| "grad_norm": 1.0899403095245361, | |
| "learning_rate": 1.7774133083411434e-05, | |
| "loss": 0.131, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 3.5660731021555763, | |
| "grad_norm": 1.66856050491333, | |
| "learning_rate": 1.7771204311152767e-05, | |
| "loss": 0.1282, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 3.570759137769447, | |
| "grad_norm": 1.9693597555160522, | |
| "learning_rate": 1.7768275538894096e-05, | |
| "loss": 0.1141, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 3.5754451733833177, | |
| "grad_norm": 2.5107643604278564, | |
| "learning_rate": 1.7765346766635426e-05, | |
| "loss": 0.1302, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 3.580131208997188, | |
| "grad_norm": 1.843782663345337, | |
| "learning_rate": 1.776241799437676e-05, | |
| "loss": 0.1295, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 3.584817244611059, | |
| "grad_norm": 1.5396898984909058, | |
| "learning_rate": 1.7759489222118088e-05, | |
| "loss": 0.1146, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 3.5895032802249296, | |
| "grad_norm": 1.250181794166565, | |
| "learning_rate": 1.775656044985942e-05, | |
| "loss": 0.0997, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 3.5941893158388005, | |
| "grad_norm": 1.0038021802902222, | |
| "learning_rate": 1.775363167760075e-05, | |
| "loss": 0.131, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 3.598875351452671, | |
| "grad_norm": 1.7689967155456543, | |
| "learning_rate": 1.775070290534208e-05, | |
| "loss": 0.1323, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 3.6035613870665415, | |
| "grad_norm": 1.6515671014785767, | |
| "learning_rate": 1.7747774133083413e-05, | |
| "loss": 0.1228, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 3.6082474226804124, | |
| "grad_norm": 1.6234405040740967, | |
| "learning_rate": 1.7744845360824743e-05, | |
| "loss": 0.1177, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 3.612933458294283, | |
| "grad_norm": 1.1468828916549683, | |
| "learning_rate": 1.7741916588566075e-05, | |
| "loss": 0.121, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 3.617619493908154, | |
| "grad_norm": 2.2512290477752686, | |
| "learning_rate": 1.7738987816307405e-05, | |
| "loss": 0.1506, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 3.6223055295220243, | |
| "grad_norm": 1.967869758605957, | |
| "learning_rate": 1.7736059044048738e-05, | |
| "loss": 0.1278, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 3.626991565135895, | |
| "grad_norm": 1.6843619346618652, | |
| "learning_rate": 1.7733130271790067e-05, | |
| "loss": 0.1292, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 3.6316776007497658, | |
| "grad_norm": 1.7128106355667114, | |
| "learning_rate": 1.77302014995314e-05, | |
| "loss": 0.1144, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 3.6316776007497658, | |
| "eval_loss": 0.043338797986507416, | |
| "eval_pearson_cosine": 0.7938926107691149, | |
| "eval_pearson_dot": 0.6518687502689318, | |
| "eval_pearson_euclidean": 0.7541829192023144, | |
| "eval_pearson_manhattan": 0.7533667893162033, | |
| "eval_runtime": 3.5292, | |
| "eval_samples_per_second": 425.028, | |
| "eval_spearman_cosine": 0.7989181032598132, | |
| "eval_spearman_dot": 0.658312239789335, | |
| "eval_spearman_euclidean": 0.7677205029787961, | |
| "eval_spearman_manhattan": 0.7673451483642694, | |
| "eval_steps_per_second": 26.635, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 3.6363636363636362, | |
| "grad_norm": 2.2347500324249268, | |
| "learning_rate": 1.772727272727273e-05, | |
| "loss": 0.145, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 3.641049671977507, | |
| "grad_norm": 1.1075607538223267, | |
| "learning_rate": 1.772434395501406e-05, | |
| "loss": 0.122, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 3.6457357075913777, | |
| "grad_norm": 1.9819576740264893, | |
| "learning_rate": 1.772141518275539e-05, | |
| "loss": 0.1542, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 3.650421743205248, | |
| "grad_norm": 1.3385450839996338, | |
| "learning_rate": 1.771848641049672e-05, | |
| "loss": 0.1269, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 3.655107778819119, | |
| "grad_norm": 1.2146624326705933, | |
| "learning_rate": 1.771555763823805e-05, | |
| "loss": 0.1124, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 3.6597938144329896, | |
| "grad_norm": 1.8543137311935425, | |
| "learning_rate": 1.771262886597938e-05, | |
| "loss": 0.1513, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 3.6644798500468605, | |
| "grad_norm": 1.147740125656128, | |
| "learning_rate": 1.7709700093720713e-05, | |
| "loss": 0.1207, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 3.669165885660731, | |
| "grad_norm": 2.1737101078033447, | |
| "learning_rate": 1.7706771321462043e-05, | |
| "loss": 0.1345, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 3.6738519212746015, | |
| "grad_norm": 2.481095790863037, | |
| "learning_rate": 1.7703842549203376e-05, | |
| "loss": 0.1304, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 3.6785379568884724, | |
| "grad_norm": 1.9203423261642456, | |
| "learning_rate": 1.7700913776944705e-05, | |
| "loss": 0.1257, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 3.683223992502343, | |
| "grad_norm": 2.324158191680908, | |
| "learning_rate": 1.7697985004686038e-05, | |
| "loss": 0.126, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 3.687910028116214, | |
| "grad_norm": 1.6233323812484741, | |
| "learning_rate": 1.7695056232427368e-05, | |
| "loss": 0.1228, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 3.6925960637300843, | |
| "grad_norm": 1.0567041635513306, | |
| "learning_rate": 1.7692127460168697e-05, | |
| "loss": 0.1129, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 3.697282099343955, | |
| "grad_norm": 1.4684457778930664, | |
| "learning_rate": 1.768919868791003e-05, | |
| "loss": 0.1164, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 3.7019681349578257, | |
| "grad_norm": 2.070171594619751, | |
| "learning_rate": 1.768626991565136e-05, | |
| "loss": 0.1488, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 3.706654170571696, | |
| "grad_norm": 1.93710458278656, | |
| "learning_rate": 1.7683341143392693e-05, | |
| "loss": 0.1423, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 3.711340206185567, | |
| "grad_norm": 2.109379529953003, | |
| "learning_rate": 1.7680412371134022e-05, | |
| "loss": 0.1247, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 3.7160262417994376, | |
| "grad_norm": 1.9402395486831665, | |
| "learning_rate": 1.7677483598875355e-05, | |
| "loss": 0.1258, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 3.720712277413308, | |
| "grad_norm": 1.3770403861999512, | |
| "learning_rate": 1.7674554826616684e-05, | |
| "loss": 0.1152, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 3.725398313027179, | |
| "grad_norm": 1.8845633268356323, | |
| "learning_rate": 1.7671626054358014e-05, | |
| "loss": 0.1128, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 3.7300843486410495, | |
| "grad_norm": 1.80978524684906, | |
| "learning_rate": 1.7668697282099347e-05, | |
| "loss": 0.1506, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 3.7347703842549205, | |
| "grad_norm": 1.714689016342163, | |
| "learning_rate": 1.7665768509840676e-05, | |
| "loss": 0.1295, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 3.739456419868791, | |
| "grad_norm": 2.2683634757995605, | |
| "learning_rate": 1.7662839737582006e-05, | |
| "loss": 0.153, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 3.7441424554826614, | |
| "grad_norm": 1.1487226486206055, | |
| "learning_rate": 1.7659910965323335e-05, | |
| "loss": 0.1305, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 3.7488284910965324, | |
| "grad_norm": 1.4769738912582397, | |
| "learning_rate": 1.7656982193064668e-05, | |
| "loss": 0.0971, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 3.7488284910965324, | |
| "eval_loss": 0.04381699115037918, | |
| "eval_pearson_cosine": 0.7952038190054296, | |
| "eval_pearson_dot": 0.6344663529067152, | |
| "eval_pearson_euclidean": 0.7547335910463318, | |
| "eval_pearson_manhattan": 0.7536726944037966, | |
| "eval_runtime": 3.6935, | |
| "eval_samples_per_second": 406.123, | |
| "eval_spearman_cosine": 0.7993143516157027, | |
| "eval_spearman_dot": 0.6382938851288388, | |
| "eval_spearman_euclidean": 0.7679077610624258, | |
| "eval_spearman_manhattan": 0.767544325158077, | |
| "eval_steps_per_second": 25.45, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 3.753514526710403, | |
| "grad_norm": 2.859675645828247, | |
| "learning_rate": 1.7654053420805998e-05, | |
| "loss": 0.1442, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 3.758200562324274, | |
| "grad_norm": 1.54740571975708, | |
| "learning_rate": 1.765112464854733e-05, | |
| "loss": 0.1194, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 3.7628865979381443, | |
| "grad_norm": 1.5696630477905273, | |
| "learning_rate": 1.764819587628866e-05, | |
| "loss": 0.1542, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 3.7675726335520148, | |
| "grad_norm": 1.8833867311477661, | |
| "learning_rate": 1.7645267104029993e-05, | |
| "loss": 0.1222, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 3.7722586691658857, | |
| "grad_norm": 2.2312562465667725, | |
| "learning_rate": 1.7642338331771322e-05, | |
| "loss": 0.1378, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 3.776944704779756, | |
| "grad_norm": 1.9470106363296509, | |
| "learning_rate": 1.7639409559512655e-05, | |
| "loss": 0.136, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 3.781630740393627, | |
| "grad_norm": 1.736140489578247, | |
| "learning_rate": 1.7636480787253985e-05, | |
| "loss": 0.144, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 3.7863167760074976, | |
| "grad_norm": 2.0280628204345703, | |
| "learning_rate": 1.7633552014995314e-05, | |
| "loss": 0.1331, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 3.791002811621368, | |
| "grad_norm": 1.1331037282943726, | |
| "learning_rate": 1.7630623242736647e-05, | |
| "loss": 0.1327, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 3.795688847235239, | |
| "grad_norm": 2.00110125541687, | |
| "learning_rate": 1.7627694470477977e-05, | |
| "loss": 0.1335, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 3.8003748828491095, | |
| "grad_norm": 1.589747428894043, | |
| "learning_rate": 1.762476569821931e-05, | |
| "loss": 0.1222, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 3.8050609184629804, | |
| "grad_norm": 1.8979542255401611, | |
| "learning_rate": 1.762183692596064e-05, | |
| "loss": 0.1534, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 3.809746954076851, | |
| "grad_norm": 2.384608268737793, | |
| "learning_rate": 1.761890815370197e-05, | |
| "loss": 0.1395, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 3.8144329896907214, | |
| "grad_norm": 1.8555763959884644, | |
| "learning_rate": 1.76159793814433e-05, | |
| "loss": 0.1298, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 3.8191190253045924, | |
| "grad_norm": 2.0810048580169678, | |
| "learning_rate": 1.761305060918463e-05, | |
| "loss": 0.1165, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 3.823805060918463, | |
| "grad_norm": 1.9116188287734985, | |
| "learning_rate": 1.761012183692596e-05, | |
| "loss": 0.1316, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 3.8284910965323338, | |
| "grad_norm": 2.3420250415802, | |
| "learning_rate": 1.7607193064667293e-05, | |
| "loss": 0.1176, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 3.8331771321462043, | |
| "grad_norm": 2.0588436126708984, | |
| "learning_rate": 1.7604264292408623e-05, | |
| "loss": 0.1365, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 3.8378631677600747, | |
| "grad_norm": 1.89832603931427, | |
| "learning_rate": 1.7601335520149952e-05, | |
| "loss": 0.1459, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 3.8425492033739457, | |
| "grad_norm": 2.0133140087127686, | |
| "learning_rate": 1.7598406747891285e-05, | |
| "loss": 0.1075, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 3.847235238987816, | |
| "grad_norm": 1.7235685586929321, | |
| "learning_rate": 1.7595477975632615e-05, | |
| "loss": 0.1143, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 3.851921274601687, | |
| "grad_norm": 1.6717524528503418, | |
| "learning_rate": 1.7592549203373948e-05, | |
| "loss": 0.1269, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 3.8566073102155576, | |
| "grad_norm": 1.5314955711364746, | |
| "learning_rate": 1.7589620431115277e-05, | |
| "loss": 0.1264, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 3.861293345829428, | |
| "grad_norm": 2.4005510807037354, | |
| "learning_rate": 1.758669165885661e-05, | |
| "loss": 0.1299, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 3.865979381443299, | |
| "grad_norm": 1.6033201217651367, | |
| "learning_rate": 1.758376288659794e-05, | |
| "loss": 0.1107, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 3.865979381443299, | |
| "eval_loss": 0.04320518299937248, | |
| "eval_pearson_cosine": 0.7953389681810563, | |
| "eval_pearson_dot": 0.6355398535557981, | |
| "eval_pearson_euclidean": 0.7518228241740452, | |
| "eval_pearson_manhattan": 0.7507054666151873, | |
| "eval_runtime": 3.6594, | |
| "eval_samples_per_second": 409.902, | |
| "eval_spearman_cosine": 0.799221389686279, | |
| "eval_spearman_dot": 0.6411224962967852, | |
| "eval_spearman_euclidean": 0.7674570737836293, | |
| "eval_spearman_manhattan": 0.7673149114040088, | |
| "eval_steps_per_second": 25.687, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 3.8706654170571695, | |
| "grad_norm": 1.8069274425506592, | |
| "learning_rate": 1.7580834114339272e-05, | |
| "loss": 0.1293, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 3.8753514526710404, | |
| "grad_norm": 1.9214448928833008, | |
| "learning_rate": 1.7577905342080602e-05, | |
| "loss": 0.1235, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 3.880037488284911, | |
| "grad_norm": 2.0332281589508057, | |
| "learning_rate": 1.757497656982193e-05, | |
| "loss": 0.1303, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 3.8847235238987814, | |
| "grad_norm": 1.8587048053741455, | |
| "learning_rate": 1.7572047797563264e-05, | |
| "loss": 0.1272, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 3.8894095595126523, | |
| "grad_norm": 1.7040314674377441, | |
| "learning_rate": 1.7569119025304594e-05, | |
| "loss": 0.1304, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 3.894095595126523, | |
| "grad_norm": 1.7979313135147095, | |
| "learning_rate": 1.7566190253045923e-05, | |
| "loss": 0.1226, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 3.8987816307403937, | |
| "grad_norm": 1.6295264959335327, | |
| "learning_rate": 1.7563261480787256e-05, | |
| "loss": 0.1198, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 3.9034676663542642, | |
| "grad_norm": 1.764858603477478, | |
| "learning_rate": 1.7560332708528586e-05, | |
| "loss": 0.1294, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 3.9081537019681347, | |
| "grad_norm": 2.3997533321380615, | |
| "learning_rate": 1.7557403936269915e-05, | |
| "loss": 0.1288, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 3.9128397375820057, | |
| "grad_norm": 2.302992820739746, | |
| "learning_rate": 1.7554475164011248e-05, | |
| "loss": 0.1482, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 3.917525773195876, | |
| "grad_norm": 1.8705153465270996, | |
| "learning_rate": 1.7551546391752578e-05, | |
| "loss": 0.1203, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 3.922211808809747, | |
| "grad_norm": 2.1490349769592285, | |
| "learning_rate": 1.754861761949391e-05, | |
| "loss": 0.1074, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 3.9268978444236176, | |
| "grad_norm": 1.5012431144714355, | |
| "learning_rate": 1.754568884723524e-05, | |
| "loss": 0.1202, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 3.931583880037488, | |
| "grad_norm": 1.2775022983551025, | |
| "learning_rate": 1.754276007497657e-05, | |
| "loss": 0.1241, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 3.936269915651359, | |
| "grad_norm": 2.361064910888672, | |
| "learning_rate": 1.7539831302717902e-05, | |
| "loss": 0.1393, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 3.9409559512652295, | |
| "grad_norm": 1.8726896047592163, | |
| "learning_rate": 1.7536902530459232e-05, | |
| "loss": 0.1191, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 3.9456419868791004, | |
| "grad_norm": 1.8277250528335571, | |
| "learning_rate": 1.7533973758200565e-05, | |
| "loss": 0.1265, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 3.950328022492971, | |
| "grad_norm": 1.534006118774414, | |
| "learning_rate": 1.7531044985941894e-05, | |
| "loss": 0.1153, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 3.9550140581068414, | |
| "grad_norm": 1.9085185527801514, | |
| "learning_rate": 1.7528116213683227e-05, | |
| "loss": 0.1216, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 3.9597000937207123, | |
| "grad_norm": 2.059544563293457, | |
| "learning_rate": 1.7525187441424557e-05, | |
| "loss": 0.1185, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 3.964386129334583, | |
| "grad_norm": 2.6308438777923584, | |
| "learning_rate": 1.752225866916589e-05, | |
| "loss": 0.1137, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 3.9690721649484537, | |
| "grad_norm": 1.2617682218551636, | |
| "learning_rate": 1.751932989690722e-05, | |
| "loss": 0.143, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 3.973758200562324, | |
| "grad_norm": 2.1921794414520264, | |
| "learning_rate": 1.751640112464855e-05, | |
| "loss": 0.1086, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 3.9784442361761947, | |
| "grad_norm": 1.7036564350128174, | |
| "learning_rate": 1.751347235238988e-05, | |
| "loss": 0.127, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 3.9831302717900656, | |
| "grad_norm": 1.6190659999847412, | |
| "learning_rate": 1.751054358013121e-05, | |
| "loss": 0.1232, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 3.9831302717900656, | |
| "eval_loss": 0.04382430762052536, | |
| "eval_pearson_cosine": 0.7946660833476784, | |
| "eval_pearson_dot": 0.6408322008008298, | |
| "eval_pearson_euclidean": 0.7506581872281402, | |
| "eval_pearson_manhattan": 0.7492952268314212, | |
| "eval_runtime": 3.2595, | |
| "eval_samples_per_second": 460.2, | |
| "eval_spearman_cosine": 0.7991836067006112, | |
| "eval_spearman_dot": 0.6471776019534929, | |
| "eval_spearman_euclidean": 0.7665561093361127, | |
| "eval_spearman_manhattan": 0.7655283844207366, | |
| "eval_steps_per_second": 28.839, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 3.987816307403936, | |
| "grad_norm": 1.5553314685821533, | |
| "learning_rate": 1.750761480787254e-05, | |
| "loss": 0.1199, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 3.992502343017807, | |
| "grad_norm": 1.8198939561843872, | |
| "learning_rate": 1.750468603561387e-05, | |
| "loss": 0.1387, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 3.9971883786316775, | |
| "grad_norm": 1.5737895965576172, | |
| "learning_rate": 1.7501757263355203e-05, | |
| "loss": 0.1291, | |
| "step": 8530 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 8536, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |