| { |
| "best_metric": 0.5882205367088318, |
| "best_model_checkpoint": "/content/My Drive/LLaMA-LoRA Tuner/lora_models/earthshine-nondecorative-2023-06-27-17-40-07/checkpoint-500", |
| "epoch": 1.404494382022472, |
| "global_step": 500, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.03, |
| "learning_rate": 2.3999999999999997e-05, |
| "loss": 2.0693, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.06, |
| "learning_rate": 5.399999999999999e-05, |
| "loss": 1.9946, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.08, |
| "learning_rate": 8.1e-05, |
| "loss": 1.8837, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.11, |
| "learning_rate": 0.00011099999999999999, |
| "loss": 1.6655, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 0.00014099999999999998, |
| "loss": 1.6095, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.17, |
| "learning_rate": 0.00017099999999999998, |
| "loss": 1.3838, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.2, |
| "learning_rate": 0.000201, |
| "loss": 1.3548, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.22, |
| "learning_rate": 0.00023099999999999998, |
| "loss": 1.0789, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.25, |
| "learning_rate": 0.000261, |
| "loss": 1.0589, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 0.00029099999999999997, |
| "loss": 1.0259, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.31, |
| "learning_rate": 0.000299393063583815, |
| "loss": 0.9936, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.34, |
| "learning_rate": 0.0002985260115606936, |
| "loss": 0.9633, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 0.00029765895953757224, |
| "loss": 1.0018, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.39, |
| "learning_rate": 0.0002967919075144509, |
| "loss": 0.893, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 0.00029592485549132945, |
| "loss": 0.8079, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.45, |
| "learning_rate": 0.0002950578034682081, |
| "loss": 0.8668, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.48, |
| "learning_rate": 0.00029419075144508666, |
| "loss": 0.773, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 0.0002933236994219653, |
| "loss": 0.8139, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.53, |
| "learning_rate": 0.0002924566473988439, |
| "loss": 0.9139, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 0.0002915895953757225, |
| "loss": 0.7371, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.59, |
| "learning_rate": 0.0002907225433526011, |
| "loss": 0.8604, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.62, |
| "learning_rate": 0.0002898554913294797, |
| "loss": 0.8597, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 0.0002889884393063584, |
| "loss": 0.7954, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.67, |
| "learning_rate": 0.00028812138728323696, |
| "loss": 0.7559, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 0.0002872543352601156, |
| "loss": 0.8488, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.73, |
| "learning_rate": 0.00028638728323699417, |
| "loss": 0.7252, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.76, |
| "learning_rate": 0.0002855202312138728, |
| "loss": 0.6751, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 0.00028465317919075143, |
| "loss": 0.7852, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.81, |
| "learning_rate": 0.00028378612716763, |
| "loss": 0.7843, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 0.00028291907514450864, |
| "loss": 0.7886, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.87, |
| "learning_rate": 0.00028205202312138727, |
| "loss": 0.7218, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.9, |
| "learning_rate": 0.0002811849710982659, |
| "loss": 0.7919, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 0.0002803179190751445, |
| "loss": 0.7248, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.96, |
| "learning_rate": 0.0002794508670520231, |
| "loss": 0.7699, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 0.00027858381502890174, |
| "loss": 0.6909, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.01, |
| "learning_rate": 0.0002777167630057803, |
| "loss": 0.6992, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.04, |
| "learning_rate": 0.00027684971098265894, |
| "loss": 0.6747, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 0.0002759826589595375, |
| "loss": 0.672, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.1, |
| "learning_rate": 0.0002751156069364162, |
| "loss": 0.5997, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.12, |
| "learning_rate": 0.0002742485549132948, |
| "loss": 0.747, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.15, |
| "learning_rate": 0.0002733815028901734, |
| "loss": 0.7217, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.18, |
| "learning_rate": 0.000272514450867052, |
| "loss": 0.683, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 0.0002716473988439306, |
| "loss": 0.63, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.24, |
| "learning_rate": 0.00027078034682080925, |
| "loss": 0.6889, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.26, |
| "learning_rate": 0.0002699132947976878, |
| "loss": 0.6582, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.29, |
| "learning_rate": 0.00026904624277456645, |
| "loss": 0.6366, |
| "step": 460 |
| }, |
| { |
| "epoch": 1.32, |
| "learning_rate": 0.00026817919075144503, |
| "loss": 0.7249, |
| "step": 470 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 0.0002673121387283237, |
| "loss": 0.53, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.38, |
| "learning_rate": 0.0002664450867052023, |
| "loss": 0.6172, |
| "step": 490 |
| }, |
| { |
| "epoch": 1.4, |
| "learning_rate": 0.0002655780346820809, |
| "loss": 0.6033, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.4, |
| "eval_loss": 0.5882205367088318, |
| "eval_runtime": 14.2208, |
| "eval_samples_per_second": 3.516, |
| "eval_steps_per_second": 0.492, |
| "step": 500 |
| } |
| ], |
| "max_steps": 3560, |
| "num_train_epochs": 10, |
| "total_flos": 1.0535163032469504e+17, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|