| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.11199211575505084, |
| "eval_steps": 1000, |
| "global_step": 5000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.002239842315101017, |
| "grad_norm": 71.10852813720703, |
| "learning_rate": 0.0001, |
| "loss": 3.5057, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.004479684630202034, |
| "grad_norm": 57.756534576416016, |
| "learning_rate": 9.9998756572327e-05, |
| "loss": 0.8254, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.006719526945303051, |
| "grad_norm": 44.61080551147461, |
| "learning_rate": 9.999502635115246e-05, |
| "loss": 0.6935, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.008959369260404068, |
| "grad_norm": 50.82619094848633, |
| "learning_rate": 9.998880952200681e-05, |
| "loss": 0.5972, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.011199211575505085, |
| "grad_norm": 44.26677703857422, |
| "learning_rate": 9.998010639409713e-05, |
| "loss": 0.5408, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.013439053890606102, |
| "grad_norm": 42.8045654296875, |
| "learning_rate": 9.996891740029186e-05, |
| "loss": 0.5509, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.01567889620570712, |
| "grad_norm": 30.989139556884766, |
| "learning_rate": 9.995524309709913e-05, |
| "loss": 0.4823, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.017918738520808136, |
| "grad_norm": 34.01952362060547, |
| "learning_rate": 9.993908416463927e-05, |
| "loss": 0.5111, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.020158580835909153, |
| "grad_norm": 34.079307556152344, |
| "learning_rate": 9.992044140661079e-05, |
| "loss": 0.4635, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.02239842315101017, |
| "grad_norm": 26.16071128845215, |
| "learning_rate": 9.989931575025056e-05, |
| "loss": 0.4883, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.02239842315101017, |
| "eval_avg_non_pair_similarity": 0.0020343252948339737, |
| "eval_avg_pair_similarity": 0.008852629057131708, |
| "eval_loss": 0.5431402921676636, |
| "eval_runtime": 19.6178, |
| "eval_samples_per_second": 25.487, |
| "eval_similarity_ratio": 4.35162905343228, |
| "eval_steps_per_second": 0.816, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.024638265466111187, |
| "grad_norm": 35.38695526123047, |
| "learning_rate": 9.987570824628759e-05, |
| "loss": 0.4655, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.026878107781212204, |
| "grad_norm": 32.071346282958984, |
| "learning_rate": 9.984962006889084e-05, |
| "loss": 0.4342, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.029117950096313218, |
| "grad_norm": 39.6610221862793, |
| "learning_rate": 9.982105251561082e-05, |
| "loss": 0.458, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.03135779241141424, |
| "grad_norm": 31.493322372436523, |
| "learning_rate": 9.979000700731491e-05, |
| "loss": 0.4525, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.03359763472651525, |
| "grad_norm": 34.453399658203125, |
| "learning_rate": 9.975648508811693e-05, |
| "loss": 0.41, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.03583747704161627, |
| "grad_norm": 33.990074157714844, |
| "learning_rate": 9.972048842530012e-05, |
| "loss": 0.4097, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.038077319356717286, |
| "grad_norm": 30.44228172302246, |
| "learning_rate": 9.968201880923439e-05, |
| "loss": 0.4257, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.040317161671818307, |
| "grad_norm": 31.427162170410156, |
| "learning_rate": 9.964107815328711e-05, |
| "loss": 0.3821, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.04255700398691932, |
| "grad_norm": 25.877887725830078, |
| "learning_rate": 9.959766849372808e-05, |
| "loss": 0.3788, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.04479684630202034, |
| "grad_norm": 25.36798095703125, |
| "learning_rate": 9.955179198962817e-05, |
| "loss": 0.3854, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.04479684630202034, |
| "eval_avg_non_pair_similarity": 0.0016289287904792565, |
| "eval_avg_pair_similarity": 0.0032495629731565715, |
| "eval_loss": 0.4394480586051941, |
| "eval_runtime": 19.5064, |
| "eval_samples_per_second": 25.633, |
| "eval_similarity_ratio": 1.9949079371360972, |
| "eval_steps_per_second": 0.82, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.047036688617121354, |
| "grad_norm": 24.537384033203125, |
| "learning_rate": 9.950345092275198e-05, |
| "loss": 0.3886, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.049276530932222375, |
| "grad_norm": 26.4981632232666, |
| "learning_rate": 9.945264769744431e-05, |
| "loss": 0.3926, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.05151637324732339, |
| "grad_norm": 40.585941314697266, |
| "learning_rate": 9.939938484051063e-05, |
| "loss": 0.3796, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.05375621556242441, |
| "grad_norm": 26.374149322509766, |
| "learning_rate": 9.934366500109132e-05, |
| "loss": 0.3723, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.05599605787752542, |
| "grad_norm": 31.54728126525879, |
| "learning_rate": 9.928549095053001e-05, |
| "loss": 0.3639, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.058235900192626436, |
| "grad_norm": 26.00455093383789, |
| "learning_rate": 9.922486558223567e-05, |
| "loss": 0.3387, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.060475742507727456, |
| "grad_norm": 23.684057235717773, |
| "learning_rate": 9.916179191153873e-05, |
| "loss": 0.3494, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.06271558482282848, |
| "grad_norm": 24.034656524658203, |
| "learning_rate": 9.909627307554108e-05, |
| "loss": 0.362, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.06495542713792948, |
| "grad_norm": 31.07891082763672, |
| "learning_rate": 9.902831233296009e-05, |
| "loss": 0.3481, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.0671952694530305, |
| "grad_norm": 24.183626174926758, |
| "learning_rate": 9.895791306396644e-05, |
| "loss": 0.3535, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.0671952694530305, |
| "eval_avg_non_pair_similarity": -0.0007878901720614375, |
| "eval_avg_pair_similarity": -0.008789425778668375, |
| "eval_loss": 0.4187028110027313, |
| "eval_runtime": 19.4839, |
| "eval_samples_per_second": 25.662, |
| "eval_similarity_ratio": 11.155648452463497, |
| "eval_steps_per_second": 0.821, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.06943511176813152, |
| "grad_norm": 20.178213119506836, |
| "learning_rate": 9.888507877001616e-05, |
| "loss": 0.3375, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.07167495408323254, |
| "grad_norm": 31.757244110107422, |
| "learning_rate": 9.880981307367627e-05, |
| "loss": 0.3489, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.07391479639833355, |
| "grad_norm": 27.889982223510742, |
| "learning_rate": 9.873211971844477e-05, |
| "loss": 0.3463, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.07615463871343457, |
| "grad_norm": 23.282093048095703, |
| "learning_rate": 9.865200256856437e-05, |
| "loss": 0.3235, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.07839448102853559, |
| "grad_norm": 24.573863983154297, |
| "learning_rate": 9.856946560883034e-05, |
| "loss": 0.3474, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.08063432334363661, |
| "grad_norm": 20.7421932220459, |
| "learning_rate": 9.848451294439224e-05, |
| "loss": 0.3533, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.08287416565873762, |
| "grad_norm": 22.051513671875, |
| "learning_rate": 9.839714880054987e-05, |
| "loss": 0.3277, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.08511400797383864, |
| "grad_norm": 22.470027923583984, |
| "learning_rate": 9.830737752254294e-05, |
| "loss": 0.3261, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.08735385028893966, |
| "grad_norm": 20.80890464782715, |
| "learning_rate": 9.821520357533513e-05, |
| "loss": 0.3474, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.08959369260404068, |
| "grad_norm": 20.187280654907227, |
| "learning_rate": 9.812063154339191e-05, |
| "loss": 0.2984, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.08959369260404068, |
| "eval_avg_non_pair_similarity": -0.0008396646829899625, |
| "eval_avg_pair_similarity": -0.011315496074035763, |
| "eval_loss": 0.32538020610809326, |
| "eval_runtime": 19.4916, |
| "eval_samples_per_second": 25.652, |
| "eval_similarity_ratio": 13.476208185560939, |
| "eval_steps_per_second": 0.821, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.09183353491914169, |
| "grad_norm": 26.496036529541016, |
| "learning_rate": 9.802366613045254e-05, |
| "loss": 0.3326, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.09407337723424271, |
| "grad_norm": 23.102359771728516, |
| "learning_rate": 9.792431215929613e-05, |
| "loss": 0.3341, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.09631321954934373, |
| "grad_norm": 21.27369499206543, |
| "learning_rate": 9.782257457150177e-05, |
| "loss": 0.3254, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.09855306186444475, |
| "grad_norm": 19.25406265258789, |
| "learning_rate": 9.771845842720274e-05, |
| "loss": 0.3334, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.10079290417954576, |
| "grad_norm": 15.666335105895996, |
| "learning_rate": 9.761196890483482e-05, |
| "loss": 0.3064, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.10303274649464678, |
| "grad_norm": 24.592592239379883, |
| "learning_rate": 9.75031113008788e-05, |
| "loss": 0.2902, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.1052725888097478, |
| "grad_norm": 20.0572566986084, |
| "learning_rate": 9.739189102959696e-05, |
| "loss": 0.3121, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.10751243112484882, |
| "grad_norm": 19.887725830078125, |
| "learning_rate": 9.727831362276381e-05, |
| "loss": 0.3014, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.10975227343994982, |
| "grad_norm": 15.561097145080566, |
| "learning_rate": 9.716238472939101e-05, |
| "loss": 0.316, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.11199211575505084, |
| "grad_norm": 17.96786880493164, |
| "learning_rate": 9.704411011544629e-05, |
| "loss": 0.3215, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.11199211575505084, |
| "eval_avg_non_pair_similarity": 0.00024146916974524413, |
| "eval_avg_pair_similarity": -0.009633154251612723, |
| "eval_loss": 0.2951599061489105, |
| "eval_runtime": 19.5025, |
| "eval_samples_per_second": 25.638, |
| "eval_similarity_ratio": -39.893930400207765, |
| "eval_steps_per_second": 0.82, |
| "step": 5000 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 44646, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 32, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|