| { |
| "best_metric": 2.4635186195373535, |
| "best_model_checkpoint": "./drive/MyDrive/peptide_esm/checkpoint-10488", |
| "epoch": 2.0, |
| "eval_steps": 500, |
| "global_step": 10488, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.1, |
| "grad_norm": 0.7004187107086182, |
| "learning_rate": 7.5e-07, |
| "loss": 2.5303, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.19, |
| "grad_norm": 0.7539237141609192, |
| "learning_rate": 1.5e-06, |
| "loss": 2.5119, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.29, |
| "grad_norm": 0.9745526313781738, |
| "learning_rate": 2.25e-06, |
| "loss": 2.4899, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.38, |
| "grad_norm": 0.6661661863327026, |
| "learning_rate": 3e-06, |
| "loss": 2.4903, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 0.7792529463768005, |
| "learning_rate": 2.9380677126341868e-06, |
| "loss": 2.4831, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.57, |
| "grad_norm": 2.211763381958008, |
| "learning_rate": 2.8762592898431047e-06, |
| "loss": 2.4785, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.67, |
| "grad_norm": 0.7369110584259033, |
| "learning_rate": 2.814450867052023e-06, |
| "loss": 2.4763, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.76, |
| "grad_norm": 0.8026860952377319, |
| "learning_rate": 2.75251857968621e-06, |
| "loss": 2.4702, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.86, |
| "grad_norm": 0.5967105627059937, |
| "learning_rate": 2.6905862923203965e-06, |
| "loss": 2.4703, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.95, |
| "grad_norm": 0.68107670545578, |
| "learning_rate": 2.6286540049545828e-06, |
| "loss": 2.4699, |
| "step": 5000 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 2.46760630607605, |
| "eval_runtime": 181.2222, |
| "eval_samples_per_second": 823.095, |
| "eval_steps_per_second": 6.434, |
| "step": 5244 |
| }, |
| { |
| "epoch": 1.05, |
| "grad_norm": 1.064468502998352, |
| "learning_rate": 2.56672171758877e-06, |
| "loss": 2.4671, |
| "step": 5500 |
| }, |
| { |
| "epoch": 1.14, |
| "grad_norm": 1.5440330505371094, |
| "learning_rate": 2.5047894302229566e-06, |
| "loss": 2.4663, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.24, |
| "grad_norm": 0.913020133972168, |
| "learning_rate": 2.442857142857143e-06, |
| "loss": 2.4654, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.33, |
| "grad_norm": 0.8316154479980469, |
| "learning_rate": 2.3810487200660612e-06, |
| "loss": 2.4694, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.43, |
| "grad_norm": 0.6185809969902039, |
| "learning_rate": 2.3191164327002475e-06, |
| "loss": 2.4668, |
| "step": 7500 |
| }, |
| { |
| "epoch": 1.53, |
| "grad_norm": 1.1215540170669556, |
| "learning_rate": 2.2571841453344346e-06, |
| "loss": 2.4668, |
| "step": 8000 |
| }, |
| { |
| "epoch": 1.62, |
| "grad_norm": 0.6679463982582092, |
| "learning_rate": 2.195251857968621e-06, |
| "loss": 2.4681, |
| "step": 8500 |
| }, |
| { |
| "epoch": 1.72, |
| "grad_norm": 1.1260002851486206, |
| "learning_rate": 2.1333195706028075e-06, |
| "loss": 2.4639, |
| "step": 9000 |
| }, |
| { |
| "epoch": 1.81, |
| "grad_norm": 0.6827268004417419, |
| "learning_rate": 2.0713872832369942e-06, |
| "loss": 2.4631, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.91, |
| "grad_norm": 0.7196531891822815, |
| "learning_rate": 2.0095788604459126e-06, |
| "loss": 2.4644, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_loss": 2.4635186195373535, |
| "eval_runtime": 180.8599, |
| "eval_samples_per_second": 824.744, |
| "eval_steps_per_second": 6.447, |
| "step": 10488 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 26220, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "total_flos": 4.1949317916310464e+17, |
| "train_batch_size": 256, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|